var/home/core/zuul-output/0000755000175000017500000000000015070121225014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070132247015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004744423715070132240017706 0ustar rootrootOct 04 04:32:44 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 04:32:44 crc restorecon[4677]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:44 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 04:32:45 crc restorecon[4677]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 04:32:50 crc kubenswrapper[4992]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 04:32:50 crc kubenswrapper[4992]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 04:32:50 crc kubenswrapper[4992]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 04:32:50 crc kubenswrapper[4992]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 04:32:50 crc kubenswrapper[4992]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 04:32:50 crc kubenswrapper[4992]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.010605 4992 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585697 4992 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585735 4992 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585740 4992 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585744 4992 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585747 4992 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585751 4992 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585755 4992 feature_gate.go:330] unrecognized feature gate: Example Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585758 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585762 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585766 4992 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585769 4992 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585773 4992 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585778 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585782 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585786 4992 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585790 4992 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585795 4992 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585798 4992 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585802 4992 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585806 4992 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585810 4992 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585814 4992 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585818 4992 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585822 4992 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585826 4992 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585830 4992 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585834 4992 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585840 4992 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585843 4992 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585847 4992 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585851 4992 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585855 4992 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585859 4992 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585863 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585867 4992 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585871 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585879 4992 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585888 4992 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585892 4992 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585900 4992 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585909 4992 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585915 4992 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585921 4992 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585928 4992 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585933 4992 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585938 4992 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585952 4992 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585956 4992 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585961 4992 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585964 4992 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585968 4992 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585972 4992 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585975 4992 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585981 4992 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585985 4992 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585989 4992 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585992 4992 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585996 4992 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.585999 4992 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586005 4992 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586009 4992 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586013 4992 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586016 4992 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586020 4992 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586023 4992 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586026 4992 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586030 4992 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586033 4992 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586036 4992 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586041 4992 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.586046 4992 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586142 4992 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586154 4992 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586180 4992 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586187 4992 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586194 4992 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586198 4992 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586205 4992 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586210 4992 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586215 4992 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586220 4992 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586226 4992 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586255 4992 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586262 4992 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586267 4992 flags.go:64] FLAG: --cgroup-root="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586272 4992 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586276 4992 flags.go:64] FLAG: --client-ca-file="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586280 4992 flags.go:64] FLAG: --cloud-config="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586284 4992 flags.go:64] FLAG: --cloud-provider="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586288 4992 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586297 4992 flags.go:64] FLAG: --cluster-domain="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586301 4992 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586305 4992 flags.go:64] FLAG: --config-dir="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586309 4992 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586314 4992 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586320 4992 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586326 4992 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586331 4992 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586337 4992 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586342 4992 flags.go:64] FLAG: --contention-profiling="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586347 4992 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586353 4992 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586379 4992 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586383 4992 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586391 4992 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586396 4992 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586401 4992 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586405 4992 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586410 4992 flags.go:64] FLAG: --enable-server="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586414 4992 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586426 4992 flags.go:64] FLAG: --event-burst="100" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586432 4992 flags.go:64] FLAG: --event-qps="50" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586437 4992 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586442 4992 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586447 4992 flags.go:64] FLAG: --eviction-hard="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586454 4992 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586460 4992 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586466 4992 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586478 4992 flags.go:64] FLAG: --eviction-soft="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586482 4992 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586486 4992 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586492 4992 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586496 4992 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586500 4992 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586504 4992 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586508 4992 flags.go:64] FLAG: --feature-gates="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586515 4992 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586521 4992 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586527 4992 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586541 4992 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586546 4992 flags.go:64] FLAG: --healthz-port="10248" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586552 4992 flags.go:64] FLAG: --help="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586557 4992 flags.go:64] FLAG: --hostname-override="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586575 4992 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586580 4992 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586586 4992 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586591 4992 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586595 4992 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586599 4992 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586603 4992 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586608 4992 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586613 4992 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586618 4992 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586624 4992 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586629 4992 flags.go:64] FLAG: --kube-reserved="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586634 4992 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586639 4992 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586645 4992 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586649 4992 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586655 4992 flags.go:64] FLAG: --lock-file="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586660 4992 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586665 4992 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586671 4992 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586680 4992 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586696 4992 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586701 4992 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586706 4992 flags.go:64] FLAG: --logging-format="text" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586711 4992 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586717 4992 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586722 4992 flags.go:64] FLAG: --manifest-url="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586727 4992 flags.go:64] FLAG: --manifest-url-header="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586735 4992 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586740 4992 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586747 4992 flags.go:64] FLAG: --max-pods="110" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586752 4992 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586760 4992 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586766 4992 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586771 4992 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586776 4992 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586781 4992 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586787 4992 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586802 4992 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586807 4992 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586812 4992 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586817 4992 flags.go:64] FLAG: --pod-cidr="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586822 4992 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586832 4992 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586837 4992 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586843 4992 flags.go:64] FLAG: --pods-per-core="0" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586848 4992 flags.go:64] FLAG: --port="10250" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586853 4992 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586858 4992 flags.go:64] FLAG: --provider-id="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586864 4992 flags.go:64] FLAG: --qos-reserved="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586870 4992 flags.go:64] FLAG: --read-only-port="10255" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586875 4992 flags.go:64] FLAG: --register-node="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586880 4992 flags.go:64] FLAG: --register-schedulable="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586885 4992 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586895 4992 flags.go:64] FLAG: --registry-burst="10" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586900 4992 flags.go:64] FLAG: --registry-qps="5" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586905 4992 flags.go:64] FLAG: --reserved-cpus="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586917 4992 flags.go:64] FLAG: --reserved-memory="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586924 4992 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586929 4992 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586935 4992 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586940 4992 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586945 4992 flags.go:64] FLAG: --runonce="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586950 4992 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586958 4992 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586963 4992 flags.go:64] FLAG: --seccomp-default="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586968 4992 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586973 4992 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586978 4992 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586984 4992 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586989 4992 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586994 4992 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.586999 4992 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587004 4992 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587009 4992 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587014 4992 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587020 4992 flags.go:64] FLAG: --system-cgroups="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587025 4992 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587034 4992 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587039 4992 flags.go:64] FLAG: --tls-cert-file="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587049 4992 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587060 4992 flags.go:64] FLAG: --tls-min-version="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587066 4992 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587071 4992 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587076 4992 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587082 4992 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587088 4992 flags.go:64] FLAG: --v="2" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587095 4992 flags.go:64] FLAG: --version="false" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587109 4992 flags.go:64] FLAG: --vmodule="" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587115 4992 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.587120 4992 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587250 4992 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587258 4992 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587275 4992 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587280 4992 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587285 4992 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587292 4992 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587297 4992 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587301 4992 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587305 4992 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587309 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587313 4992 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587317 4992 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587321 4992 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587325 4992 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587329 4992 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587348 4992 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587371 4992 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587376 4992 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587380 4992 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587383 4992 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587387 4992 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587392 4992 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587396 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587400 4992 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587404 4992 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587409 4992 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587419 4992 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587425 4992 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587430 4992 feature_gate.go:330] unrecognized feature gate: Example Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587434 4992 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587439 4992 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587443 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587448 4992 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587452 4992 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587456 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587461 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587465 4992 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587473 4992 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587484 4992 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587488 4992 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587492 4992 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587497 4992 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587502 4992 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587506 4992 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587510 4992 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587514 4992 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587518 4992 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587521 4992 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587525 4992 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587528 4992 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587532 4992 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587535 4992 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587539 4992 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587543 4992 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587546 4992 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587549 4992 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587553 4992 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587556 4992 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587561 4992 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587564 4992 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587568 4992 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587572 4992 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587577 4992 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587582 4992 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587587 4992 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587590 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587594 4992 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587598 4992 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587601 4992 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587608 4992 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.587611 4992 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.588501 4992 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.597585 4992 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.597626 4992 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597697 4992 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597705 4992 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597711 4992 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597719 4992 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597723 4992 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597728 4992 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597732 4992 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597737 4992 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597741 4992 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597746 4992 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597751 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597756 4992 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597761 4992 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597766 4992 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597771 4992 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597776 4992 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597781 4992 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597786 4992 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597791 4992 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597796 4992 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597799 4992 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597803 4992 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597807 4992 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597811 4992 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597814 4992 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597818 4992 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597822 4992 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597826 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597831 4992 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597836 4992 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597840 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597844 4992 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597848 4992 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597853 4992 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597863 4992 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597868 4992 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597887 4992 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597892 4992 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597896 4992 feature_gate.go:330] unrecognized feature gate: Example Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597902 4992 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597906 4992 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597910 4992 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597915 4992 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597920 4992 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597925 4992 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597930 4992 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597935 4992 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597939 4992 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597944 4992 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597949 4992 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597953 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597957 4992 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597961 4992 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597965 4992 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597969 4992 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597973 4992 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597976 4992 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597980 4992 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597984 4992 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597988 4992 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597992 4992 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597996 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.597999 4992 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598003 4992 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598007 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598011 4992 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598014 4992 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598018 4992 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598022 4992 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598026 4992 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598031 4992 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.598038 4992 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598172 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598179 4992 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598183 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598188 4992 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598191 4992 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598195 4992 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598199 4992 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598204 4992 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598208 4992 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598212 4992 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598216 4992 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598220 4992 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598224 4992 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598227 4992 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598231 4992 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598235 4992 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598239 4992 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598244 4992 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598248 4992 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598252 4992 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598256 4992 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598261 4992 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598266 4992 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598271 4992 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598277 4992 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598281 4992 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598285 4992 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598288 4992 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598292 4992 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598296 4992 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598300 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598304 4992 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598307 4992 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598311 4992 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598316 4992 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598320 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598323 4992 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598327 4992 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598331 4992 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598334 4992 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598338 4992 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598343 4992 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598347 4992 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598351 4992 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598355 4992 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598378 4992 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598383 4992 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598388 4992 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598393 4992 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598398 4992 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598403 4992 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598408 4992 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598414 4992 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598419 4992 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598423 4992 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598428 4992 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598432 4992 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598437 4992 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598441 4992 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598446 4992 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598450 4992 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598454 4992 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598459 4992 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598464 4992 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598468 4992 feature_gate.go:330] unrecognized feature gate: Example Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598473 4992 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598479 4992 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598484 4992 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598489 4992 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598493 4992 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 04:32:50 crc kubenswrapper[4992]: W1004 04:32:50.598499 4992 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.598506 4992 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.598701 4992 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.800644 4992 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.800767 4992 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.805777 4992 server.go:997] "Starting client certificate rotation" Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.805813 4992 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.806020 4992 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-14 15:09:24.153037465 +0000 UTC Oct 04 04:32:50 crc kubenswrapper[4992]: I1004 04:32:50.806134 4992 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2458h36m33.346906831s for next certificate rotation Oct 04 04:32:51 crc kubenswrapper[4992]: I1004 04:32:51.072865 4992 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 04:32:51 crc kubenswrapper[4992]: I1004 04:32:51.088435 4992 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 04:32:51 crc kubenswrapper[4992]: I1004 04:32:51.120082 4992 log.go:25] "Validated CRI v1 runtime API" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.177660 4992 log.go:25] "Validated CRI v1 image API" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.181302 4992 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.188286 4992 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-04-27-55-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.188330 4992 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.204844 4992 manager.go:217] Machine: {Timestamp:2025-10-04 04:32:54.202024976 +0000 UTC m=+8.049700464 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:e9142861-7515-4077-bd2a-8f44fe92e099 BootID:0b6f8901-3900-4293-9527-60030bed2cef Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:97:30:cc Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:97:30:cc Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9e:fd:0e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:98:41:75 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:2e:06:3e Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b8:45:97 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2e:1a:7c:58:f8:ce Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:66:d1:6d:60:e5:77 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.205105 4992 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.205307 4992 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.208776 4992 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.209010 4992 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.209046 4992 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.209763 4992 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.209789 4992 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.210324 4992 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.210353 4992 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.210590 4992 state_mem.go:36] "Initialized new in-memory state store" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.210677 4992 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.225743 4992 kubelet.go:418] "Attempting to sync node with API server" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.225829 4992 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.225870 4992 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.225900 4992 kubelet.go:324] "Adding apiserver pod source" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.225927 4992 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.230805 4992 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.232139 4992 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 04:32:54 crc kubenswrapper[4992]: W1004 04:32:54.233031 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.233207 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:54 crc kubenswrapper[4992]: W1004 04:32:54.233044 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.233295 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.244305 4992 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247143 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247176 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247186 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247196 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247212 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247227 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247238 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247255 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247264 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247273 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247287 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.247296 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.249786 4992 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.250240 4992 server.go:1280] "Started kubelet" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.250451 4992 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.250856 4992 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.251378 4992 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.251797 4992 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 04:32:54 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.253617 4992 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.253665 4992 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.253708 4992 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 03:43:17.947533354 +0000 UTC Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.253779 4992 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1439h10m23.693758677s for next certificate rotation Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.254012 4992 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.255674 4992 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.255711 4992 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.255804 4992 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.255764 4992 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="200ms" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.256544 4992 server.go:460] "Adding debug handlers to kubelet server" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.256605 4992 factory.go:55] Registering systemd factory Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.256633 4992 factory.go:221] Registration of the systemd container factory successfully Oct 04 04:32:54 crc kubenswrapper[4992]: W1004 04:32:54.256997 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.257081 4992 factory.go:153] Registering CRI-O factory Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.257081 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.257110 4992 factory.go:221] Registration of the crio container factory successfully Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.257218 4992 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.257250 4992 factory.go:103] Registering Raw factory Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.257271 4992 manager.go:1196] Started watching for new ooms in manager Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.257882 4992 manager.go:319] Starting recovery of all containers Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.270114 4992 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.238:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b2f8fd6e0f9b6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 04:32:54.25020767 +0000 UTC m=+8.097883138,LastTimestamp:2025-10-04 04:32:54.25020767 +0000 UTC m=+8.097883138,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.287935 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288054 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288073 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288088 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288104 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288118 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288130 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288143 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288161 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288176 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288189 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288203 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288218 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288243 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288258 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288273 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288293 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288308 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288322 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288335 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288347 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288381 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288397 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288409 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288422 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288434 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288453 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288466 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.288478 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290491 4992 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290556 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290582 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290603 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290620 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290635 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290653 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290671 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290685 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290707 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290721 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290736 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290785 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290798 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290841 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290858 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290875 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290889 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290935 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290956 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290979 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.290993 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291008 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291021 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291043 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291060 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291074 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291089 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291103 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291115 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291161 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291179 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291192 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291206 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291220 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291239 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291254 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291269 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291282 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291294 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291307 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291322 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291336 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291351 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291389 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291406 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291419 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291434 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291445 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291459 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291473 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291488 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291502 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291516 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291533 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291546 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291566 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291581 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291624 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291641 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291656 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291672 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291688 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291702 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291719 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291735 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291748 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291765 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291779 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291789 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291800 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291812 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291825 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291838 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291850 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291863 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291882 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291895 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291910 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291925 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291942 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291960 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291981 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.291999 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292012 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292025 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292039 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292050 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292067 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292080 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292092 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292105 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292118 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292132 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292145 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292156 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292167 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292181 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292196 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292238 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292251 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292262 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292274 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292285 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292296 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292308 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292320 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292351 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292385 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292399 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292411 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292425 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292436 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292448 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292459 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292497 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292548 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292575 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292589 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292603 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292616 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292627 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292639 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292677 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292691 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292743 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292758 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292770 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292814 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292828 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292842 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292857 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292875 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292888 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292900 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292914 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292927 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292940 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292953 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.292991 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293005 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293017 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293029 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293043 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293062 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293074 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293087 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293101 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293112 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293124 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293137 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293150 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293170 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293183 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293194 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293229 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293242 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293255 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293267 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293281 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293294 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293309 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293321 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293339 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293354 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293530 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293545 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293562 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293579 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293592 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293605 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293619 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293632 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293645 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293657 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293669 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293714 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293728 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293741 4992 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293756 4992 reconstruct.go:97] "Volume reconstruction finished" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.293767 4992 reconciler.go:26] "Reconciler: start to sync state" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.295587 4992 manager.go:324] Recovery completed Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.306936 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.309053 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.309101 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.309111 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.310461 4992 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.310490 4992 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.310515 4992 state_mem.go:36] "Initialized new in-memory state store" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.315670 4992 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.317857 4992 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.317909 4992 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.317941 4992 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.318094 4992 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 04:32:54 crc kubenswrapper[4992]: W1004 04:32:54.319369 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.319552 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.354459 4992 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.354779 4992 policy_none.go:49] "None policy: Start" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.355695 4992 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.355725 4992 state_mem.go:35] "Initializing new in-memory state store" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.418430 4992 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.442557 4992 manager.go:334] "Starting Device Plugin manager" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.442607 4992 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.442620 4992 server.go:79] "Starting device plugin registration server" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.443037 4992 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.443052 4992 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.449101 4992 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.449332 4992 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.449457 4992 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.449471 4992 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.456815 4992 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="400ms" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.543656 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.545383 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.545425 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.545441 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.545478 4992 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.546012 4992 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.619200 4992 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.619629 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.620782 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.620814 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.620823 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.620935 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.621686 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.621704 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.621714 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.627282 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.627319 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.627347 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.627324 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.627290 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.628171 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.628204 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.628222 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629007 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629034 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629127 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629106 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629210 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629269 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629291 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629388 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.629420 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.630077 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.630097 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.630106 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.630733 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.630763 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.630776 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.630902 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.631012 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.631055 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.631783 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.631822 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.631830 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.631892 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.631918 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.631932 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.632252 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.632296 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.633166 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.633183 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.633192 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.699631 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.699700 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.699808 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.699868 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.699898 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700259 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700581 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700603 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700690 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700764 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700841 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700879 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700906 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700925 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.700963 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.746899 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.748633 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.748719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.748734 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.748796 4992 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.749744 4992 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802265 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802332 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802406 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802432 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802454 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802480 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802503 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802523 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802549 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802563 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802586 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802608 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802628 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802645 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802649 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802666 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802681 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802682 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802673 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802713 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802755 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802755 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802775 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802777 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802788 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802807 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802818 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802842 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.803207 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.802562 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: E1004 04:32:54.858260 4992 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="800ms" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.966627 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:32:54 crc kubenswrapper[4992]: I1004 04:32:54.996273 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.017094 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.025016 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.029184 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.088742 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-fed04fda4a726cbc9e38762a7ae1c40b05dd76bc2e66db9dd92fcdcae36256ba WatchSource:0}: Error finding container fed04fda4a726cbc9e38762a7ae1c40b05dd76bc2e66db9dd92fcdcae36256ba: Status 404 returned error can't find the container with id fed04fda4a726cbc9e38762a7ae1c40b05dd76bc2e66db9dd92fcdcae36256ba Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.095235 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-777a5e901f91c2564e3e01f31128693643b6d32ed0c926953bcc87902264239e WatchSource:0}: Error finding container 777a5e901f91c2564e3e01f31128693643b6d32ed0c926953bcc87902264239e: Status 404 returned error can't find the container with id 777a5e901f91c2564e3e01f31128693643b6d32ed0c926953bcc87902264239e Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.097908 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-3165b7fba8ac4a398c7511943eaad4ca43286fdbec6123579f412f027ca47803 WatchSource:0}: Error finding container 3165b7fba8ac4a398c7511943eaad4ca43286fdbec6123579f412f027ca47803: Status 404 returned error can't find the container with id 3165b7fba8ac4a398c7511943eaad4ca43286fdbec6123579f412f027ca47803 Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.102701 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0cabb79e9a48be27fc7cab5ce9f1ea4a930cf868f7c312f8a74d0cc06e362e0f WatchSource:0}: Error finding container 0cabb79e9a48be27fc7cab5ce9f1ea4a930cf868f7c312f8a74d0cc06e362e0f: Status 404 returned error can't find the container with id 0cabb79e9a48be27fc7cab5ce9f1ea4a930cf868f7c312f8a74d0cc06e362e0f Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.103308 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-212a10a9d3cbe0edda9f4c2c0ea49cf3ed5f6002cb25c53974ac98b3d24c91ad WatchSource:0}: Error finding container 212a10a9d3cbe0edda9f4c2c0ea49cf3ed5f6002cb25c53974ac98b3d24c91ad: Status 404 returned error can't find the container with id 212a10a9d3cbe0edda9f4c2c0ea49cf3ed5f6002cb25c53974ac98b3d24c91ad Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.150280 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.151593 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.151636 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.151651 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.151679 4992 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:32:55 crc kubenswrapper[4992]: E1004 04:32:55.152281 4992 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.252446 4992 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.252827 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:55 crc kubenswrapper[4992]: E1004 04:32:55.252974 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.322484 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0cabb79e9a48be27fc7cab5ce9f1ea4a930cf868f7c312f8a74d0cc06e362e0f"} Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.324134 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3165b7fba8ac4a398c7511943eaad4ca43286fdbec6123579f412f027ca47803"} Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.325054 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.325140 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"777a5e901f91c2564e3e01f31128693643b6d32ed0c926953bcc87902264239e"} Oct 04 04:32:55 crc kubenswrapper[4992]: E1004 04:32:55.325140 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.326567 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fed04fda4a726cbc9e38762a7ae1c40b05dd76bc2e66db9dd92fcdcae36256ba"} Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.327572 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"212a10a9d3cbe0edda9f4c2c0ea49cf3ed5f6002cb25c53974ac98b3d24c91ad"} Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.332429 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:55 crc kubenswrapper[4992]: E1004 04:32:55.332507 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:55 crc kubenswrapper[4992]: W1004 04:32:55.513023 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:55 crc kubenswrapper[4992]: E1004 04:32:55.513120 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:55 crc kubenswrapper[4992]: E1004 04:32:55.659741 4992 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="1.6s" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.952696 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.954061 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.954134 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.954147 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:55 crc kubenswrapper[4992]: I1004 04:32:55.954179 4992 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:32:55 crc kubenswrapper[4992]: E1004 04:32:55.954975 4992 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Oct 04 04:32:56 crc kubenswrapper[4992]: I1004 04:32:56.252885 4992 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.252351 4992 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:57 crc kubenswrapper[4992]: E1004 04:32:57.260915 4992 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="3.2s" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.333206 4992 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc" exitCode=0 Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.333385 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.333569 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc"} Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.337140 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.337400 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.337415 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.339353 4992 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7" exitCode=0 Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.339454 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7"} Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.339570 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.340659 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.340705 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.340719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.341553 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6"} Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.341588 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857"} Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.342308 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.342911 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.342962 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.342976 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.343086 4992 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6" exitCode=0 Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.343172 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6"} Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.343185 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.344449 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.344484 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.344500 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.344619 4992 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="efa25998cda3065a1e3634dfc72257edd52cc51ea13c97460b2d5e322a0a6e40" exitCode=0 Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.344646 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"efa25998cda3065a1e3634dfc72257edd52cc51ea13c97460b2d5e322a0a6e40"} Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.344694 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.345401 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.345441 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.345456 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:57 crc kubenswrapper[4992]: W1004 04:32:57.395276 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:57 crc kubenswrapper[4992]: E1004 04:32:57.395379 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.555423 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.557484 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.557532 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.557546 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:57 crc kubenswrapper[4992]: I1004 04:32:57.557574 4992 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:32:57 crc kubenswrapper[4992]: E1004 04:32:57.558405 4992 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Oct 04 04:32:57 crc kubenswrapper[4992]: W1004 04:32:57.917203 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:57 crc kubenswrapper[4992]: E1004 04:32:57.917299 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.253147 4992 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.349589 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"919162be30623317a3e81e1435503638a2c7ae724ca46c797213c34cefe590c0"} Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.349648 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.350891 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.350921 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.350932 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.351734 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752"} Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.351766 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce"} Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.357521 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f"} Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.357575 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524"} Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.359092 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f"} Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.359147 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d"} Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.359253 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:58 crc kubenswrapper[4992]: W1004 04:32:58.359962 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:58 crc kubenswrapper[4992]: E1004 04:32:58.360038 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.360285 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.360385 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.360397 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.363872 4992 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea" exitCode=0 Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.363916 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea"} Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.364040 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.364876 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.364915 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:58 crc kubenswrapper[4992]: I1004 04:32:58.364927 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:58 crc kubenswrapper[4992]: W1004 04:32:58.411594 4992 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:58 crc kubenswrapper[4992]: E1004 04:32:58.411842 4992 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.253330 4992 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.370574 4992 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad" exitCode=0 Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.370652 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad"} Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.370669 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.372282 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.372324 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.372335 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.375121 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03"} Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.375330 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.376613 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.376750 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.376827 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.380285 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a"} Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.380332 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109"} Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.380348 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55"} Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.380378 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.380385 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.380687 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.381234 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.381274 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.381286 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.382261 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.382299 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.382309 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.382341 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.382406 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:32:59 crc kubenswrapper[4992]: I1004 04:32:59.382418 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.022149 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.155563 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.385871 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.385928 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.386439 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351"} Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.386472 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053"} Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.386482 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506"} Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.386491 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a"} Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.386569 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.386587 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.386920 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.388075 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.388100 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.388111 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.388775 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.388787 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.388794 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.389148 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.389158 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.389166 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.759477 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.761413 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.761484 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.761499 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:00 crc kubenswrapper[4992]: I1004 04:33:00.761535 4992 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.337088 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.391754 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.391805 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.392258 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.392507 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6"} Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.392588 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.392939 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.392970 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.392979 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.392946 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.393118 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.393133 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.393659 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.393684 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:01 crc kubenswrapper[4992]: I1004 04:33:01.393692 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.393812 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.395141 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.395254 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.395330 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.783794 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.783996 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.785988 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.786072 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.786097 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.823182 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.823407 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.824577 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.824644 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.824657 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.975164 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.975438 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.980283 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.980324 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:02 crc kubenswrapper[4992]: I1004 04:33:02.980334 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:03 crc kubenswrapper[4992]: I1004 04:33:03.022471 4992 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded" start-of-body= Oct 04 04:33:03 crc kubenswrapper[4992]: I1004 04:33:03.022586 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.156433 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.156639 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.157937 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.158113 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.158416 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:04 crc kubenswrapper[4992]: E1004 04:33:04.449252 4992 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.932823 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.933037 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.934183 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.934210 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.934219 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:04 crc kubenswrapper[4992]: I1004 04:33:04.938515 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.402138 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.403009 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.403049 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.403060 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.670097 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.670326 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.671492 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.671531 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:05 crc kubenswrapper[4992]: I1004 04:33:05.671543 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:06 crc kubenswrapper[4992]: I1004 04:33:06.734211 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 04:33:06 crc kubenswrapper[4992]: I1004 04:33:06.734483 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:06 crc kubenswrapper[4992]: I1004 04:33:06.735644 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:06 crc kubenswrapper[4992]: I1004 04:33:06.735708 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:06 crc kubenswrapper[4992]: I1004 04:33:06.735723 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:09 crc kubenswrapper[4992]: I1004 04:33:09.794038 4992 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 04:33:09 crc kubenswrapper[4992]: I1004 04:33:09.794112 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 04:33:09 crc kubenswrapper[4992]: I1004 04:33:09.799385 4992 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 04:33:09 crc kubenswrapper[4992]: I1004 04:33:09.799459 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 04:33:10 crc kubenswrapper[4992]: I1004 04:33:10.159336 4992 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]log ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]etcd ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/priority-and-fairness-filter ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-apiextensions-informers ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-apiextensions-controllers ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/crd-informer-synced ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-system-namespaces-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 04 04:33:10 crc kubenswrapper[4992]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 04 04:33:10 crc kubenswrapper[4992]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/bootstrap-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/start-kube-aggregator-informers ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/apiservice-registration-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/apiservice-discovery-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]autoregister-completion ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/apiservice-openapi-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 04 04:33:10 crc kubenswrapper[4992]: livez check failed Oct 04 04:33:10 crc kubenswrapper[4992]: I1004 04:33:10.159424 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:33:11 crc kubenswrapper[4992]: I1004 04:33:11.341380 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:33:11 crc kubenswrapper[4992]: I1004 04:33:11.341515 4992 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 04:33:11 crc kubenswrapper[4992]: I1004 04:33:11.342613 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:11 crc kubenswrapper[4992]: I1004 04:33:11.342660 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:11 crc kubenswrapper[4992]: I1004 04:33:11.342672 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:13 crc kubenswrapper[4992]: I1004 04:33:13.022780 4992 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 04:33:13 crc kubenswrapper[4992]: I1004 04:33:13.022868 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 04:33:14 crc kubenswrapper[4992]: E1004 04:33:14.449351 4992 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 04:33:14 crc kubenswrapper[4992]: E1004 04:33:14.793702 4992 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.802460 4992 trace.go:236] Trace[428483661]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 04:33:01.301) (total time: 13500ms): Oct 04 04:33:14 crc kubenswrapper[4992]: Trace[428483661]: ---"Objects listed" error: 13500ms (04:33:14.802) Oct 04 04:33:14 crc kubenswrapper[4992]: Trace[428483661]: [13.500507387s] [13.500507387s] END Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.802537 4992 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.834860 4992 trace.go:236] Trace[1732542173]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 04:33:03.462) (total time: 11372ms): Oct 04 04:33:14 crc kubenswrapper[4992]: Trace[1732542173]: ---"Objects listed" error: 11372ms (04:33:14.834) Oct 04 04:33:14 crc kubenswrapper[4992]: Trace[1732542173]: [11.372412634s] [11.372412634s] END Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.834895 4992 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.834860 4992 trace.go:236] Trace[170363904]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 04:33:01.737) (total time: 13097ms): Oct 04 04:33:14 crc kubenswrapper[4992]: Trace[170363904]: ---"Objects listed" error: 13097ms (04:33:14.834) Oct 04 04:33:14 crc kubenswrapper[4992]: Trace[170363904]: [13.097394159s] [13.097394159s] END Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.834947 4992 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.835969 4992 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.838527 4992 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.838739 4992 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.839702 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.839731 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.839744 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.839787 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.839801 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:14Z","lastTransitionTime":"2025-10-04T04:33:14Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.840026 4992 trace.go:236] Trace[747719140]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 04:33:02.649) (total time: 12190ms): Oct 04 04:33:14 crc kubenswrapper[4992]: Trace[747719140]: ---"Objects listed" error: 12190ms (04:33:14.839) Oct 04 04:33:14 crc kubenswrapper[4992]: Trace[747719140]: [12.190448392s] [12.190448392s] END Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.840053 4992 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.851339 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:14 crc kubenswrapper[4992]: E1004 04:33:14.859270 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.862682 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.862756 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.862768 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.862787 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.862801 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:14Z","lastTransitionTime":"2025-10-04T04:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:14 crc kubenswrapper[4992]: E1004 04:33:14.879119 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.914466 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.914530 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.914543 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.914562 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.914573 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:14Z","lastTransitionTime":"2025-10-04T04:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:14 crc kubenswrapper[4992]: E1004 04:33:14.932189 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.944027 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.944070 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.944080 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.944098 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.944110 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:14Z","lastTransitionTime":"2025-10-04T04:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:14 crc kubenswrapper[4992]: E1004 04:33:14.959755 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.962888 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.962923 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.962931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.962945 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.962954 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:14Z","lastTransitionTime":"2025-10-04T04:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:14 crc kubenswrapper[4992]: E1004 04:33:14.972634 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:14 crc kubenswrapper[4992]: E1004 04:33:14.972756 4992 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.974426 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.974456 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.974464 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.974481 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:14 crc kubenswrapper[4992]: I1004 04:33:14.974492 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:14Z","lastTransitionTime":"2025-10-04T04:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.077914 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.077959 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.077969 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.077998 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.078009 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.159417 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.180323 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.180391 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.180403 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.180420 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.180437 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.241240 4992 apiserver.go:52] "Watching apiserver" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.244443 4992 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.245080 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-ncbdl","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.246236 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ncbdl" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.246642 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.247025 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.247147 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.247150 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.247216 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.247420 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.247375 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.249271 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.249511 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.251418 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.252159 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.252985 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.254566 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.254866 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.254939 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.254974 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.254995 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.255014 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.255467 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.255599 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.257199 4992 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.258929 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.281007 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.282231 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.282267 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.282282 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.282302 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.282315 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.293997 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.306622 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.316277 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.339744 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.339798 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.339827 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.339848 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.339910 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.339935 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.339955 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.339980 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340005 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340045 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340070 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340096 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340116 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340140 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340162 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340269 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340292 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340317 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340627 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340656 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340677 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340698 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340717 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340736 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340758 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340781 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340800 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340819 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340839 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340861 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340882 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340914 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340937 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340957 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.340981 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341010 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341034 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341066 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341089 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341111 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341132 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341159 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341181 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341199 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341219 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341246 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341272 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341291 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341311 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341330 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341350 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341388 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341411 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341430 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341453 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341472 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341493 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341512 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341534 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341576 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341599 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341622 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341645 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341670 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341694 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341717 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341741 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341764 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341799 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341820 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341841 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341862 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341884 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341905 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341928 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341949 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341970 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.341993 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342016 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342038 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342061 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342081 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342101 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342126 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342147 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342168 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342192 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342215 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342236 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342258 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342278 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342300 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342321 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342344 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342384 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342409 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342432 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342456 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342477 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342499 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342520 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342542 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342565 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342588 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342609 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342633 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342655 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342676 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342696 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342718 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342738 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342758 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342780 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342829 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342852 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342873 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342895 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342916 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342936 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342958 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342979 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.342999 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343020 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343041 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343064 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343087 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343120 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343143 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343166 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343190 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343213 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343235 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343260 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343283 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343305 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343328 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343353 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343393 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343416 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343438 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343460 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343483 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343506 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343530 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343553 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343575 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343597 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343621 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343644 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343666 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343690 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343715 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343739 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343761 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343784 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343810 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343836 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343860 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343887 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343912 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343935 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343958 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.343980 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344002 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344029 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344052 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344074 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344098 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344123 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344148 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344173 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344198 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344223 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344247 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344268 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344290 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344288 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344313 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344438 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344470 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344497 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344521 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344544 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344577 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344606 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344634 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344654 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344673 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344690 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344713 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344781 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344801 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344819 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344837 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344855 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344874 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344919 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344941 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344983 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.344978 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345006 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345027 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345051 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345082 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345111 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345139 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345170 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345199 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345232 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345261 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345286 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345326 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ad0f7c37-e53f-4a34-8da9-575e78b32a95-hosts-file\") pod \"node-resolver-ncbdl\" (UID: \"ad0f7c37-e53f-4a34-8da9-575e78b32a95\") " pod="openshift-dns/node-resolver-ncbdl" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345375 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f9fp\" (UniqueName: \"kubernetes.io/projected/ad0f7c37-e53f-4a34-8da9-575e78b32a95-kube-api-access-9f9fp\") pod \"node-resolver-ncbdl\" (UID: \"ad0f7c37-e53f-4a34-8da9-575e78b32a95\") " pod="openshift-dns/node-resolver-ncbdl" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345464 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345478 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.347037 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.379419 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384564 4992 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345199 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345780 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345804 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345804 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.345960 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.346012 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.346548 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.346583 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.346710 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.346722 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.346738 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.346958 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.347013 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.347051 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.347076 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.347092 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.347149 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.348343 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:33:15.848318183 +0000 UTC m=+29.695993721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388264 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388478 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388544 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388771 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388811 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388854 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388872 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.389048 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.390385 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.390697 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.390867 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.391245 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.391098 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.391259 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.391323 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.391387 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.391824 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.392489 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.392716 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.394463 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.396050 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.396584 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.397192 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.391578 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.397860 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.398177 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.398287 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.398493 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388672 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.399296 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.399352 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.399463 4992 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.399464 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.399559 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:15.899535091 +0000 UTC m=+29.747210559 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.400192 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.400086 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.351772 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.351827 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.351852 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.352083 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.352292 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.352325 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.400353 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.368505 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.373090 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.376388 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.376664 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.400498 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.376740 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.400523 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.377108 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.377130 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.377165 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.377488 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.377752 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.377774 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.377852 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378167 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378240 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378308 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378270 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378384 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378501 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378541 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378573 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378648 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378400 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.378516 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.379718 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.400704 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.381982 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384088 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384208 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384352 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384503 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384717 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384778 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.400780 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384873 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384813 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.384915 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.385056 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.385883 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.385941 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.386115 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.386435 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.385504 4992 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.400932 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:15.900913775 +0000 UTC m=+29.748589243 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.385669 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.386398 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.387556 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.387601 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.387787 4992 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.401016 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:15.901009088 +0000 UTC m=+29.748684556 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.388003 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.401154 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.401754 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.401832 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.401855 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.402240 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.351553 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.402346 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.402548 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.414574 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.414621 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.414651 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.402552 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.402590 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.414839 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.414869 4992 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.403143 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.403182 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.403499 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.403713 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.403793 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.403881 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.403934 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.404177 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.404320 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.404477 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.404539 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.405702 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.412409 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.404746 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.404987 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.405205 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.405927 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.406732 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.408443 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.408887 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.408960 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.409128 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.412866 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.413046 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.413143 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.408725 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.413172 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.415131 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:15.915110273 +0000 UTC m=+29.762785741 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.415657 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.418226 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.418256 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.418566 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.418660 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.418981 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.419141 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.423225 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.423299 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.423972 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.429707 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.429896 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.430028 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.430084 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.430433 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.430694 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.431055 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.431224 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.432795 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.433708 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.433778 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.433820 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.434123 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.434267 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.434320 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.435128 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.435151 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.435210 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.435242 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.435352 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.435679 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.435788 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.435801 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.436208 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.436649 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.436692 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.436797 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.436952 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.437020 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.437159 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.437244 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.437396 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.437538 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.437955 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.438287 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.438523 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.438775 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.438805 4992 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a" exitCode=255 Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.438824 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.438843 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.438869 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.439487 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.439519 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.439897 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.440072 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.440094 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.440308 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.440711 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.441052 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.441098 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.441327 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.441716 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.442090 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.443611 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.447782 4992 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.448890 4992 scope.go:117] "RemoveContainer" containerID="a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.448912 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.448949 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ad0f7c37-e53f-4a34-8da9-575e78b32a95-hosts-file\") pod \"node-resolver-ncbdl\" (UID: \"ad0f7c37-e53f-4a34-8da9-575e78b32a95\") " pod="openshift-dns/node-resolver-ncbdl" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.448969 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.448992 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f9fp\" (UniqueName: \"kubernetes.io/projected/ad0f7c37-e53f-4a34-8da9-575e78b32a95-kube-api-access-9f9fp\") pod \"node-resolver-ncbdl\" (UID: \"ad0f7c37-e53f-4a34-8da9-575e78b32a95\") " pod="openshift-dns/node-resolver-ncbdl" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449042 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449053 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449063 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449074 4992 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449084 4992 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449200 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449234 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449465 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449517 4992 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449537 4992 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449551 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449566 4992 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449579 4992 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449624 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449633 4992 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449518 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/ad0f7c37-e53f-4a34-8da9-575e78b32a95-hosts-file\") pod \"node-resolver-ncbdl\" (UID: \"ad0f7c37-e53f-4a34-8da9-575e78b32a95\") " pod="openshift-dns/node-resolver-ncbdl" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449643 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449673 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449685 4992 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449694 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449726 4992 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449737 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449745 4992 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449753 4992 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449762 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449772 4992 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449781 4992 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449811 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449820 4992 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449830 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449839 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449850 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449859 4992 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449888 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449914 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449924 4992 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449933 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449961 4992 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449972 4992 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.449989 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450001 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450012 4992 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450048 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450062 4992 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450072 4992 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450081 4992 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450092 4992 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450102 4992 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450171 4992 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450181 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450218 4992 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450227 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450236 4992 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450246 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450293 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450304 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450315 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450324 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450376 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450482 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450492 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450567 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450577 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450586 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450619 4992 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450627 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450639 4992 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450649 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450660 4992 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.450694 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451405 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451476 4992 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451487 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451496 4992 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451557 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451602 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451640 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451654 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451664 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451675 4992 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452071 4992 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452085 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452112 4992 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452120 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452129 4992 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452137 4992 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.451950 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452145 4992 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452392 4992 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452404 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452412 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452421 4992 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452432 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452441 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452451 4992 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452460 4992 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452469 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452477 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452489 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452498 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452507 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452516 4992 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452525 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452536 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452545 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452554 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452563 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452572 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452581 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452590 4992 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452598 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452607 4992 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452617 4992 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452626 4992 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452635 4992 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452644 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452653 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452662 4992 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452671 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452680 4992 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452714 4992 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452724 4992 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452734 4992 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452742 4992 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452751 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452760 4992 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452799 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452808 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452816 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452824 4992 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452835 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452844 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452876 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452885 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452893 4992 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452901 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452909 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452917 4992 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452944 4992 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452954 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452963 4992 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452971 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452979 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452988 4992 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.452995 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453003 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453031 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453039 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453048 4992 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453057 4992 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453065 4992 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453074 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453106 4992 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453117 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453125 4992 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453133 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453141 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453150 4992 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453159 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453189 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453198 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453207 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453216 4992 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453225 4992 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453234 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453292 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453302 4992 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453311 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453321 4992 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453352 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453389 4992 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453398 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453407 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453415 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453424 4992 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453433 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453461 4992 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453470 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453479 4992 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453497 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453507 4992 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453515 4992 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453544 4992 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453553 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453564 4992 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453574 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453582 4992 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453592 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453640 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453650 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453660 4992 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453669 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453679 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.453065 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.460734 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.463128 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f9fp\" (UniqueName: \"kubernetes.io/projected/ad0f7c37-e53f-4a34-8da9-575e78b32a95-kube-api-access-9f9fp\") pod \"node-resolver-ncbdl\" (UID: \"ad0f7c37-e53f-4a34-8da9-575e78b32a95\") " pod="openshift-dns/node-resolver-ncbdl" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.469015 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.476395 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.477985 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.478286 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.488391 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.499655 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.518346 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.520204 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.520231 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.520239 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.520255 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.521700 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.527397 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.532085 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.547934 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.558264 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.558542 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.558620 4992 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.558849 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.566465 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ncbdl" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.566725 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.574297 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.584965 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.592581 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 04:33:15 crc kubenswrapper[4992]: W1004 04:33:15.601658 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-6e0cafa6866a422d1d5389502a0dfd221257a2f295895afc6baba2df942a599d WatchSource:0}: Error finding container 6e0cafa6866a422d1d5389502a0dfd221257a2f295895afc6baba2df942a599d: Status 404 returned error can't find the container with id 6e0cafa6866a422d1d5389502a0dfd221257a2f295895afc6baba2df942a599d Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.623944 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.623988 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.623999 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.624018 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.624030 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.727666 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.727717 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.727728 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.727744 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.727755 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.830113 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.830148 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.830158 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.830175 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.830187 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.860599 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.860942 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:33:16.860906394 +0000 UTC m=+30.708581862 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.903344 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-bckw6"] Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.903750 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.905803 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.906024 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.906330 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.907067 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.928197 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.932208 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.932248 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.932259 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.932279 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.932292 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:15Z","lastTransitionTime":"2025-10-04T04:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.936555 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.950485 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.961635 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.961678 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.961700 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.961729 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/361f0cb6-df75-48b8-916b-4f6e19bd8834-serviceca\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.961751 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.961773 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdh7c\" (UniqueName: \"kubernetes.io/projected/361f0cb6-df75-48b8-916b-4f6e19bd8834-kube-api-access-hdh7c\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.961791 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/361f0cb6-df75-48b8-916b-4f6e19bd8834-host\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.961967 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962003 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962017 4992 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962068 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:16.962053527 +0000 UTC m=+30.809729035 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962130 4992 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962158 4992 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962191 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962230 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962242 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:16.962219552 +0000 UTC m=+30.809895080 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962244 4992 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962269 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:16.962259943 +0000 UTC m=+30.809935441 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:15 crc kubenswrapper[4992]: E1004 04:33:15.962389 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:16.962345025 +0000 UTC m=+30.810020513 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.964186 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.975349 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.988448 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:15 crc kubenswrapper[4992]: I1004 04:33:15.998964 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.008230 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.019455 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.034996 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.035030 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.035039 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.035054 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.035063 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.062522 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdh7c\" (UniqueName: \"kubernetes.io/projected/361f0cb6-df75-48b8-916b-4f6e19bd8834-kube-api-access-hdh7c\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.062570 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/361f0cb6-df75-48b8-916b-4f6e19bd8834-host\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.062617 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/361f0cb6-df75-48b8-916b-4f6e19bd8834-serviceca\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.062735 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/361f0cb6-df75-48b8-916b-4f6e19bd8834-host\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.063693 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/361f0cb6-df75-48b8-916b-4f6e19bd8834-serviceca\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.079902 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdh7c\" (UniqueName: \"kubernetes.io/projected/361f0cb6-df75-48b8-916b-4f6e19bd8834-kube-api-access-hdh7c\") pod \"node-ca-bckw6\" (UID: \"361f0cb6-df75-48b8-916b-4f6e19bd8834\") " pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.138293 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.138338 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.138347 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.138381 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.138392 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.217834 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bckw6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.240781 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.240820 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.240831 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.240849 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.240860 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.278727 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-vhq64"] Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.279342 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.284205 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.284885 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.284977 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.287479 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.287523 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.296303 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.311580 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.322506 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.322661 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.329759 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.330576 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.330551 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.332102 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.333073 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.334445 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.343453 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.343792 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.343849 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.343867 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.343894 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.343914 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.344398 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.346521 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.347288 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.350957 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.351786 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.355847 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.360546 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.361737 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.362379 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.364345 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.365060 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.365755 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-multus-certs\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.366191 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-cnibin\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.366218 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-cni-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.366263 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-system-cni-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.366289 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-k8s-cni-cncf-io\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.366329 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-kubelet\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.366978 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j29w\" (UniqueName: \"kubernetes.io/projected/2b6879c5-46f4-4d87-959b-03dff08cff4c-kube-api-access-9j29w\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367073 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b6879c5-46f4-4d87-959b-03dff08cff4c-cni-binary-copy\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367141 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-netns\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367169 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-cni-multus\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367197 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-hostroot\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367225 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-daemon-config\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367253 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-etc-kubernetes\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367333 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-os-release\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367429 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-socket-dir-parent\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367498 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-cni-bin\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.367523 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-conf-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.369902 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.370437 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.372034 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.372628 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.372953 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.373546 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.376300 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.376961 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.379292 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.379806 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.381203 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.382155 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.382799 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.384971 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.385476 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.387163 4992 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.387484 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.387802 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.391177 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.392457 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.393042 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.394797 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.396093 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.396992 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.397652 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.398922 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.399516 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.400509 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.401576 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.402244 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.403148 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.403714 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.404654 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.405347 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.405971 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.406828 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.408106 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.409273 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.410206 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.410697 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.413671 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.443644 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.443700 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fa1df662618ea000bb5b5fa163ac995fa9ffc77e4e98dbfe066e1795d5c67972"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.445649 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.445678 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.445688 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.445705 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.445718 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.446420 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bckw6" event={"ID":"361f0cb6-df75-48b8-916b-4f6e19bd8834","Type":"ContainerStarted","Data":"0fd287f9f1f9bb09775c628d42dad482d960e27f75fc25123da28608e7811b55"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.451040 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.451104 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.451117 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7eb5f423ded9768fda2e88fbccdf4e00f63b3b1a7bdff4897dc3f29e34b91127"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.452177 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6e0cafa6866a422d1d5389502a0dfd221257a2f295895afc6baba2df942a599d"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.454328 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.457166 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ncbdl" event={"ID":"ad0f7c37-e53f-4a34-8da9-575e78b32a95","Type":"ContainerStarted","Data":"9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.457234 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ncbdl" event={"ID":"ad0f7c37-e53f-4a34-8da9-575e78b32a95","Type":"ContainerStarted","Data":"8723a177ee8b0ddd1c24da7e38df006942f5381aa3d2472782e038b5bc52214b"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.462015 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468072 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b6879c5-46f4-4d87-959b-03dff08cff4c-cni-binary-copy\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468308 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-netns\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468328 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-cni-multus\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468348 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-kubelet\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468389 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j29w\" (UniqueName: \"kubernetes.io/projected/2b6879c5-46f4-4d87-959b-03dff08cff4c-kube-api-access-9j29w\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468423 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-os-release\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468446 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-socket-dir-parent\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468461 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-hostroot\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468475 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-daemon-config\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468490 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-etc-kubernetes\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468538 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-cni-bin\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468559 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-conf-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468595 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-multus-certs\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468633 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-cnibin\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468661 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-cni-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468691 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-system-cni-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468711 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-k8s-cni-cncf-io\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.468778 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-k8s-cni-cncf-io\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469216 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-os-release\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469296 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-socket-dir-parent\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469483 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2b6879c5-46f4-4d87-959b-03dff08cff4c-cni-binary-copy\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469514 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-cni-bin\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469533 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-cni-multus\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469531 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-etc-kubernetes\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469569 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-conf-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469560 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-hostroot\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469550 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-netns\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469737 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-cnibin\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469777 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-run-multus-certs\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469874 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-system-cni-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.469889 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-cni-dir\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.470174 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/2b6879c5-46f4-4d87-959b-03dff08cff4c-multus-daemon-config\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.470199 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b6879c5-46f4-4d87-959b-03dff08cff4c-host-var-lib-kubelet\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.470972 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.471502 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.484187 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.507895 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.508079 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j29w\" (UniqueName: \"kubernetes.io/projected/2b6879c5-46f4-4d87-959b-03dff08cff4c-kube-api-access-9j29w\") pod \"multus-vhq64\" (UID: \"2b6879c5-46f4-4d87-959b-03dff08cff4c\") " pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.527666 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.545153 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.548331 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.548354 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.548377 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.548392 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.548402 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.560505 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.578305 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.595026 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.600639 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vhq64" Oct 04 04:33:16 crc kubenswrapper[4992]: W1004 04:33:16.614589 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b6879c5_46f4_4d87_959b_03dff08cff4c.slice/crio-fc32db3b6a0b1c69e959a86d95d11fa5719c59da70ca10b4cf6f3b787c454c6b WatchSource:0}: Error finding container fc32db3b6a0b1c69e959a86d95d11fa5719c59da70ca10b4cf6f3b787c454c6b: Status 404 returned error can't find the container with id fc32db3b6a0b1c69e959a86d95d11fa5719c59da70ca10b4cf6f3b787c454c6b Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.635883 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.649882 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.649922 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.649934 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.649952 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.649965 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.661458 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-wp9cz"] Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.662306 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.662522 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.663345 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-p5l9q"] Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.663950 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.665168 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.665662 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.665788 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.666966 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.667003 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.668010 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670373 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d6d4bf54-7f00-4c33-830c-8368dd9678df-proxy-tls\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670405 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5c79\" (UniqueName: \"kubernetes.io/projected/d6d4bf54-7f00-4c33-830c-8368dd9678df-kube-api-access-q5c79\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670440 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cni-binary-copy\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670526 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b2p7\" (UniqueName: \"kubernetes.io/projected/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-kube-api-access-7b2p7\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670559 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-system-cni-dir\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670593 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cnibin\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670616 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-os-release\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670634 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d6d4bf54-7f00-4c33-830c-8368dd9678df-mcd-auth-proxy-config\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670669 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670689 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d6d4bf54-7f00-4c33-830c-8368dd9678df-rootfs\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.670755 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.672928 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.687751 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.710812 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.726981 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.745932 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.752295 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.752343 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.752368 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.752389 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.752401 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.762002 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.770099 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771267 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771307 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d6d4bf54-7f00-4c33-830c-8368dd9678df-rootfs\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771334 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d6d4bf54-7f00-4c33-830c-8368dd9678df-proxy-tls\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771379 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5c79\" (UniqueName: \"kubernetes.io/projected/d6d4bf54-7f00-4c33-830c-8368dd9678df-kube-api-access-q5c79\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771463 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cni-binary-copy\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771487 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b2p7\" (UniqueName: \"kubernetes.io/projected/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-kube-api-access-7b2p7\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771521 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-system-cni-dir\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771557 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-os-release\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771504 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d6d4bf54-7f00-4c33-830c-8368dd9678df-rootfs\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771581 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cnibin\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771732 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-system-cni-dir\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771675 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-os-release\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771645 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cnibin\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771841 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d6d4bf54-7f00-4c33-830c-8368dd9678df-mcd-auth-proxy-config\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.771957 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.772084 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.772895 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.774268 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-cni-binary-copy\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.778059 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d6d4bf54-7f00-4c33-830c-8368dd9678df-mcd-auth-proxy-config\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.783961 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d6d4bf54-7f00-4c33-830c-8368dd9678df-proxy-tls\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.786918 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.790913 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.791687 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b2p7\" (UniqueName: \"kubernetes.io/projected/be8a9bf6-a143-43ab-8f47-f3a72a2e902f-kube-api-access-7b2p7\") pod \"multus-additional-cni-plugins-wp9cz\" (UID: \"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\") " pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.792239 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.793731 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5c79\" (UniqueName: \"kubernetes.io/projected/d6d4bf54-7f00-4c33-830c-8368dd9678df-kube-api-access-q5c79\") pod \"machine-config-daemon-p5l9q\" (UID: \"d6d4bf54-7f00-4c33-830c-8368dd9678df\") " pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.818530 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.834089 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.851304 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.855154 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.855224 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.855239 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.855260 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.855275 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.866277 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.873439 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.873637 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:33:18.873595262 +0000 UTC m=+32.721270730 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.882405 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.899089 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.922089 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.936536 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.951654 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.958614 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.958800 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.958823 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.958912 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.958935 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:16Z","lastTransitionTime":"2025-10-04T04:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.966085 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.975020 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.975077 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.975297 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.975323 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975444 4992 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975481 4992 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975509 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:18.975493965 +0000 UTC m=+32.823169423 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975493 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975590 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975614 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975642 4992 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975634 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975679 4992 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975568 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:18.975546856 +0000 UTC m=+32.823222324 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975746 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:18.975734101 +0000 UTC m=+32.823409669 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:16 crc kubenswrapper[4992]: E1004 04:33:16.975760 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:18.975753621 +0000 UTC m=+32.823429229 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.983092 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.992265 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:16 crc kubenswrapper[4992]: I1004 04:33:16.993403 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.001707 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:33:17 crc kubenswrapper[4992]: W1004 04:33:17.003916 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe8a9bf6_a143_43ab_8f47_f3a72a2e902f.slice/crio-2dc4c4cbeeb34c4364903369fe6b6116634ef26e3e58bf936a2c1a95bff67453 WatchSource:0}: Error finding container 2dc4c4cbeeb34c4364903369fe6b6116634ef26e3e58bf936a2c1a95bff67453: Status 404 returned error can't find the container with id 2dc4c4cbeeb34c4364903369fe6b6116634ef26e3e58bf936a2c1a95bff67453 Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.006316 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: W1004 04:33:17.012898 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6d4bf54_7f00_4c33_830c_8368dd9678df.slice/crio-626771073bb1e1023d49c3519b7a37b9b8a35855fd14ef8f655e3c6ddc84b53b WatchSource:0}: Error finding container 626771073bb1e1023d49c3519b7a37b9b8a35855fd14ef8f655e3c6ddc84b53b: Status 404 returned error can't find the container with id 626771073bb1e1023d49c3519b7a37b9b8a35855fd14ef8f655e3c6ddc84b53b Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.020046 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.035062 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.053067 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.069633 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fnr76"] Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.070239 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.070291 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.070303 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.070325 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.070342 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.070770 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.074737 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.075969 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.076161 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.076337 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.076751 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.077036 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.077123 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.077295 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.092157 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.107143 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.133060 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.147120 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.164075 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.173287 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.173556 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.173658 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.173820 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.173892 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.178457 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-ovn\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.178793 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-ovn-kubernetes\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.178904 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-config\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179026 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-script-lib\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179109 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179182 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-slash\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179284 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-node-log\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179381 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-var-lib-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179459 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-log-socket\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179532 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-netns\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179509 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179661 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-bin\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179711 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-netd\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179734 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovn-node-metrics-cert\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179808 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-env-overrides\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179877 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-systemd\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179901 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-etc-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179946 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-kubelet\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.179983 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-systemd-units\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.180047 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.180087 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7qm9\" (UniqueName: \"kubernetes.io/projected/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-kube-api-access-b7qm9\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.214512 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.251524 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.276977 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.277018 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.277031 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.277050 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.277063 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281565 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-ovn\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281657 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-ovn-kubernetes\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281677 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-config\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281675 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-ovn\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281705 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-script-lib\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281766 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-ovn-kubernetes\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281798 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281773 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281832 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-slash\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281874 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-node-log\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281901 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-var-lib-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281922 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-log-socket\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281944 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-netns\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281972 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-bin\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.281993 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-netd\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282018 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovn-node-metrics-cert\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282048 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-systemd\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282072 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-etc-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282088 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-env-overrides\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282104 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-systemd-units\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282121 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282144 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-kubelet\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282162 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7qm9\" (UniqueName: \"kubernetes.io/projected/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-kube-api-access-b7qm9\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282320 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-slash\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282345 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-node-log\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282390 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-var-lib-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282414 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-log-socket\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282434 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-netns\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282454 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-bin\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282474 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-netd\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282523 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-script-lib\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282591 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-config\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282636 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282660 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-systemd-units\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282683 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-kubelet\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282704 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-systemd\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282724 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-etc-openvswitch\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.282990 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-env-overrides\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.287297 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovn-node-metrics-cert\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.292673 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.318464 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:17 crc kubenswrapper[4992]: E1004 04:33:17.318630 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.318715 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:17 crc kubenswrapper[4992]: E1004 04:33:17.318783 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.322934 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7qm9\" (UniqueName: \"kubernetes.io/projected/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-kube-api-access-b7qm9\") pod \"ovnkube-node-fnr76\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.358537 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.379470 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.379509 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.379518 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.379532 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.379542 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.400503 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.407936 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:17 crc kubenswrapper[4992]: W1004 04:33:17.425844 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e301ed8_9506_4d6e_bbce_d2ad3817ca36.slice/crio-2976e96f9c7962bd0e54bc0651abe8f495b61bf46ef033a09c179306afea1e9c WatchSource:0}: Error finding container 2976e96f9c7962bd0e54bc0651abe8f495b61bf46ef033a09c179306afea1e9c: Status 404 returned error can't find the container with id 2976e96f9c7962bd0e54bc0651abe8f495b61bf46ef033a09c179306afea1e9c Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.435891 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.475944 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.479753 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bckw6" event={"ID":"361f0cb6-df75-48b8-916b-4f6e19bd8834","Type":"ContainerStarted","Data":"9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.480947 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"2976e96f9c7962bd0e54bc0651abe8f495b61bf46ef033a09c179306afea1e9c"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.480985 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.481011 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.481030 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.481046 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.481055 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.482158 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhq64" event={"ID":"2b6879c5-46f4-4d87-959b-03dff08cff4c","Type":"ContainerStarted","Data":"7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.482185 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhq64" event={"ID":"2b6879c5-46f4-4d87-959b-03dff08cff4c","Type":"ContainerStarted","Data":"fc32db3b6a0b1c69e959a86d95d11fa5719c59da70ca10b4cf6f3b787c454c6b"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.483779 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerStarted","Data":"8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.483800 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerStarted","Data":"2dc4c4cbeeb34c4364903369fe6b6116634ef26e3e58bf936a2c1a95bff67453"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.486151 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.486170 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.486179 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"626771073bb1e1023d49c3519b7a37b9b8a35855fd14ef8f655e3c6ddc84b53b"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.513076 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.556810 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.584308 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.584381 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.584393 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.584410 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.584422 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.600715 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.633436 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.678725 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.688214 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.689424 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.689438 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.689456 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.689469 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.713645 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.752039 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.792882 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.792924 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.792933 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.792953 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.792963 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.796671 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.835446 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.878603 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.896043 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.896099 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.896111 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.896128 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.896139 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.916095 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.953632 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.994858 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.998516 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.998558 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.998568 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.998583 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:17 crc kubenswrapper[4992]: I1004 04:33:17.998595 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:17Z","lastTransitionTime":"2025-10-04T04:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.034271 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.072954 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.101335 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.101400 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.101412 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.101439 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.101452 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.113528 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.152040 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.194698 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.204103 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.204143 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.204154 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.204174 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.204185 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.239015 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.273498 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.306421 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.306455 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.306469 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.306500 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.306521 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.319198 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:18 crc kubenswrapper[4992]: E1004 04:33:18.319393 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.408945 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.408997 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.409010 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.409028 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.409039 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.489978 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3" exitCode=0 Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.490109 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.492566 4992 generic.go:334] "Generic (PLEG): container finished" podID="be8a9bf6-a143-43ab-8f47-f3a72a2e902f" containerID="8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9" exitCode=0 Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.492670 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerDied","Data":"8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.497560 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.506113 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.511146 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.511184 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.511196 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.511229 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.511242 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.518564 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.539185 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.555631 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.574580 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.591723 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.604546 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.614563 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.614607 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.614618 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.614637 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.614648 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.619826 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.632139 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.672611 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.711995 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.716721 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.716781 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.716799 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.716826 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.716845 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.759056 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.795469 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.820494 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.820565 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.820577 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.820593 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.820603 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.831315 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.869817 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.898524 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:18 crc kubenswrapper[4992]: E1004 04:33:18.898711 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:33:22.89867695 +0000 UTC m=+36.746352418 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.914168 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.925313 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.925396 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.925416 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.925449 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.925470 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:18Z","lastTransitionTime":"2025-10-04T04:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.959868 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:18 crc kubenswrapper[4992]: I1004 04:33:18.997929 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.000165 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.000322 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.000490 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.000626 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.000429 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.000857 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.000962 4992 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.001090 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:23.001074505 +0000 UTC m=+36.848749973 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.000497 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.001285 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.001402 4992 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.001511 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:23.001502025 +0000 UTC m=+36.849177493 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.000538 4992 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.000723 4992 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.001942 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:23.001918296 +0000 UTC m=+36.849593764 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.002148 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:23.002136961 +0000 UTC m=+36.849812429 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.028523 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.028835 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.028907 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.028970 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.029028 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.031863 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.071684 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.110796 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.131628 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.131664 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.131674 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.131692 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.131704 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.150342 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.197501 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.231543 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.234920 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.234964 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.234977 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.234998 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.235017 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.275160 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.314678 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.318773 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.318914 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.319023 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:19 crc kubenswrapper[4992]: E1004 04:33:19.319222 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.337737 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.338085 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.338184 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.338278 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.338419 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.351508 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.393742 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.441310 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.441376 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.441389 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.441410 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.441465 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.503874 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.503945 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.503963 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.505755 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerStarted","Data":"738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.523673 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.543669 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.543714 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.543725 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.543743 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.543780 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.553902 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.568558 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.585595 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.607391 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.631584 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.646119 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.646166 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.646196 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.646212 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.646222 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.680148 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.713613 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.748539 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.748589 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.748600 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.748617 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.748632 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.755307 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.810844 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.850771 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.851153 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.851175 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.851185 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.851202 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.851213 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.875662 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.913188 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.952391 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:19Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.953423 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.953462 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.953471 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.953486 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:19 crc kubenswrapper[4992]: I1004 04:33:19.953497 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:19Z","lastTransitionTime":"2025-10-04T04:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.025485 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.029197 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.033626 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.040497 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.056401 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.056442 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.056451 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.056467 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.056496 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.064413 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.095275 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.133632 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.158791 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.158831 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.158840 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.158856 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.158867 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.172530 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.214632 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.257475 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.261342 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.261401 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.261413 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.261464 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.261477 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.290997 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.318885 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:20 crc kubenswrapper[4992]: E1004 04:33:20.319070 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.335615 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.363751 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.363799 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.363811 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.363829 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.363842 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.375984 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.414521 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.454301 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.465986 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.466205 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.466310 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.466407 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.466489 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.494183 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.510246 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.510286 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.532679 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.568879 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.568920 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.568929 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.568945 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.568957 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.573380 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.610739 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.659114 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.671857 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.671905 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.671917 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.671935 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.671949 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.695313 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.736524 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.776257 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.776520 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.776537 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.776559 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.776573 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.781352 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.812179 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.859784 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.879796 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.879861 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.879878 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.879897 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.879910 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.894885 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.930291 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.973800 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:20Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.982671 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.982711 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.982722 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.982740 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:20 crc kubenswrapper[4992]: I1004 04:33:20.982754 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:20Z","lastTransitionTime":"2025-10-04T04:33:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.012302 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.057515 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.085087 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.085126 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.085137 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.085157 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.085169 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.094013 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.132444 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.187919 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.187965 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.187976 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.187992 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.188003 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.291149 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.291209 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.291224 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.291261 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.291277 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.318797 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:21 crc kubenswrapper[4992]: E1004 04:33:21.318946 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.318998 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:21 crc kubenswrapper[4992]: E1004 04:33:21.319053 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.393919 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.393956 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.393967 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.393986 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.393998 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.496263 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.496638 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.496805 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.497006 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.497209 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.515778 4992 generic.go:334] "Generic (PLEG): container finished" podID="be8a9bf6-a143-43ab-8f47-f3a72a2e902f" containerID="738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084" exitCode=0 Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.515899 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerDied","Data":"738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.521366 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.536855 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.551988 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.566435 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.593878 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.600239 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.600302 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.600314 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.600331 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.600343 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.609791 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.623495 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.637620 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.652231 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.671382 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.685906 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.697257 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.703354 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.703417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.703429 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.703448 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.703458 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.711302 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.724014 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.736955 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.747507 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:21Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.806342 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.806401 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.806413 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.806430 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.806443 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.909026 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.909082 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.909099 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.909125 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:21 crc kubenswrapper[4992]: I1004 04:33:21.909143 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:21Z","lastTransitionTime":"2025-10-04T04:33:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.012087 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.012137 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.012148 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.012168 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.012180 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.116372 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.116438 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.116455 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.116474 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.116486 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.220439 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.220480 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.220490 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.220506 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.220517 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.323140 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.323194 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.323205 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.323220 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.323233 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.356135 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:22 crc kubenswrapper[4992]: E1004 04:33:22.356301 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.426945 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.426996 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.427011 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.427030 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.427058 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.529220 4992 generic.go:334] "Generic (PLEG): container finished" podID="be8a9bf6-a143-43ab-8f47-f3a72a2e902f" containerID="20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4" exitCode=0 Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.529288 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerDied","Data":"20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.530364 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.530425 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.530438 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.530455 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.530466 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.553949 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.571797 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.587420 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.615702 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.630522 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.633861 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.633898 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.633910 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.633926 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.633939 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.645997 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.659147 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.670514 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.689956 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.702780 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.715372 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.728432 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.736425 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.736477 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.736494 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.736518 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.736535 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.743897 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.760395 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.772096 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:22Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.839466 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.839503 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.839515 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.839532 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.839545 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.940094 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:22 crc kubenswrapper[4992]: E1004 04:33:22.940479 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:33:30.940432204 +0000 UTC m=+44.788107672 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.941586 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.941679 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.941698 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.941721 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:22 crc kubenswrapper[4992]: I1004 04:33:22.941735 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:22Z","lastTransitionTime":"2025-10-04T04:33:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.041637 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.041719 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.041756 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.041794 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.041895 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.041956 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.041971 4992 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.041971 4992 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.041919 4992 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.042047 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:31.042022619 +0000 UTC m=+44.889698097 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.042115 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:31.042089701 +0000 UTC m=+44.889765369 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.042132 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:31.042124422 +0000 UTC m=+44.889800130 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.042276 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.042298 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.042311 4992 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.042453 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:31.042428049 +0000 UTC m=+44.890103517 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.045122 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.045192 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.045207 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.045231 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.045249 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.148132 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.148173 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.148182 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.148200 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.148212 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.251864 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.251929 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.251946 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.251967 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.251980 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.318937 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.318980 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.319163 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:23 crc kubenswrapper[4992]: E1004 04:33:23.319298 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.358737 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.358788 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.358801 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.358881 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.358920 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.462348 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.462428 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.462442 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.462461 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.462473 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.537115 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.539603 4992 generic.go:334] "Generic (PLEG): container finished" podID="be8a9bf6-a143-43ab-8f47-f3a72a2e902f" containerID="2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99" exitCode=0 Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.539637 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerDied","Data":"2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.553071 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.566914 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.566960 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.566968 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.566988 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.566999 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.568208 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.588103 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.609333 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.625723 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.643392 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.660102 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.669602 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.669654 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.669668 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.669689 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.669706 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.677539 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.694177 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.711264 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.725563 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.739330 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.750448 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.765728 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.772882 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.772944 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.772958 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.772980 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.772994 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.789575 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:23Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.876076 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.876133 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.876149 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.876170 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.876181 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.979350 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.979425 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.979435 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.979450 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:23 crc kubenswrapper[4992]: I1004 04:33:23.979463 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:23Z","lastTransitionTime":"2025-10-04T04:33:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.082058 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.082104 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.082112 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.082130 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.082141 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.185558 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.185604 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.185612 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.185630 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.185643 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.289086 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.289129 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.289141 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.289157 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.289171 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.318757 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:24 crc kubenswrapper[4992]: E1004 04:33:24.318963 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.333933 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.345957 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.359542 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.373817 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.387890 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.392847 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.392915 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.392924 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.392942 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.392959 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.401040 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.421788 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.442248 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.456674 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.469330 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.483732 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.495266 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.495340 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.495398 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.495424 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.495442 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.503455 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.514335 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.523778 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.535170 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.545189 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerStarted","Data":"cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.557995 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.569491 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.589690 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.598160 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.598193 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.598203 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.598218 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.598228 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.605230 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.615995 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.630361 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.642870 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.657629 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.671786 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.684016 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.696697 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.700676 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.700709 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.700720 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.700736 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.700747 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.715124 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.738400 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.753241 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.767947 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:24Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.803106 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.803149 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.803163 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.803183 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.803194 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.906699 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.906747 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.906766 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.906785 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:24 crc kubenswrapper[4992]: I1004 04:33:24.906798 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:24Z","lastTransitionTime":"2025-10-04T04:33:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.009931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.009993 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.010019 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.010046 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.010062 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.112635 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.112680 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.112690 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.112703 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.112713 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.213705 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.213781 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.213793 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.213812 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.213827 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: E1004 04:33:25.228691 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.234037 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.234096 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.234109 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.234129 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.234143 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: E1004 04:33:25.246399 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.251699 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.251752 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.251763 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.251780 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.251790 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: E1004 04:33:25.265485 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.269467 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.269506 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.269516 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.269534 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.269547 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: E1004 04:33:25.290080 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.315991 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.316027 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.316054 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.316072 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.316082 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.318486 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.318535 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:25 crc kubenswrapper[4992]: E1004 04:33:25.318636 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:25 crc kubenswrapper[4992]: E1004 04:33:25.318799 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:25 crc kubenswrapper[4992]: E1004 04:33:25.337049 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: E1004 04:33:25.337171 4992 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.338943 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.338972 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.338983 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.338999 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.339012 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.441868 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.441934 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.441947 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.441964 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.441976 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.544818 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.544862 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.544872 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.544890 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.544904 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.550539 4992 generic.go:334] "Generic (PLEG): container finished" podID="be8a9bf6-a143-43ab-8f47-f3a72a2e902f" containerID="cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6" exitCode=0 Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.550577 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerDied","Data":"cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.569842 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.587557 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.606606 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.618555 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.633491 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.646744 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.646779 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.646790 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.646808 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.646820 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.649940 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.661804 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.677748 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.691418 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.703259 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.718210 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.733296 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.749796 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.749817 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.749825 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.749839 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.749848 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.752851 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.767323 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.782407 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:25Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.852563 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.852607 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.852619 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.852635 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.852648 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.954605 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.954640 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.954648 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.954663 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:25 crc kubenswrapper[4992]: I1004 04:33:25.954672 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:25Z","lastTransitionTime":"2025-10-04T04:33:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.057662 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.057708 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.057719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.057736 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.057753 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.159948 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.160010 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.160033 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.160058 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.160074 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.263165 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.263254 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.263267 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.263288 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.263302 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.318539 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:26 crc kubenswrapper[4992]: E1004 04:33:26.318758 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.365793 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.365838 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.365849 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.365867 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.365879 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.467801 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.467849 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.467899 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.467930 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.467943 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.558967 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.560132 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.561111 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.566444 4992 generic.go:334] "Generic (PLEG): container finished" podID="be8a9bf6-a143-43ab-8f47-f3a72a2e902f" containerID="5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460" exitCode=0 Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.566478 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerDied","Data":"5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.569478 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.569504 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.569512 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.569524 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.569533 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.586314 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.586848 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.588101 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.604845 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.620031 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.633416 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.644628 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.657864 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.672280 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.672480 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.672611 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.672636 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.672656 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.672670 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.684123 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.698710 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.717251 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.740293 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.753655 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.765191 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.780389 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.780714 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.780745 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.780754 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.780770 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.780788 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.792531 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.807005 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.826646 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.841465 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.853222 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.865084 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.874248 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.883301 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.883329 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.883337 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.883350 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.883358 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.896988 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.911217 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.922159 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.933894 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.950921 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.968179 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.982387 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.986115 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.986141 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.986152 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.986169 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.986181 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:26Z","lastTransitionTime":"2025-10-04T04:33:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:26 crc kubenswrapper[4992]: I1004 04:33:26.996577 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:26Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.009491 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.089308 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.089351 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.089387 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.089403 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.089414 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.192219 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.192292 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.192308 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.192333 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.192348 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.294716 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.294769 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.294785 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.294806 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.294827 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.318747 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.318757 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:27 crc kubenswrapper[4992]: E1004 04:33:27.319001 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:27 crc kubenswrapper[4992]: E1004 04:33:27.319090 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.397560 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.397613 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.397627 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.397647 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.397661 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.500134 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.500175 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.500183 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.500197 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.500208 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.575517 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" event={"ID":"be8a9bf6-a143-43ab-8f47-f3a72a2e902f","Type":"ContainerStarted","Data":"d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.575634 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.592087 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.603043 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.603094 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.603120 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.603137 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.603148 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.611999 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.626003 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.637667 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.652834 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.676075 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.690152 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.705954 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.706018 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.706031 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.706048 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.706062 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.706840 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.723316 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.736030 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.751062 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.773257 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.787628 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.804231 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.807966 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.807999 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.808011 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.808027 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.808040 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.819230 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:27Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.910632 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.910685 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.910700 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.910720 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:27 crc kubenswrapper[4992]: I1004 04:33:27.910733 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:27Z","lastTransitionTime":"2025-10-04T04:33:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.013249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.013298 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.013310 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.013330 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.013345 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.116062 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.116140 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.116194 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.116229 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.116248 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.218943 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.218986 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.218996 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.219010 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.219022 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.318847 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:28 crc kubenswrapper[4992]: E1004 04:33:28.319000 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.321583 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.321624 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.321633 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.321649 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.321659 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.423560 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.423600 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.423620 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.423636 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.423646 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.526386 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.526418 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.526426 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.526442 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.526451 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.578404 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.586813 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9"] Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.587237 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.590980 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.591157 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.599594 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.609457 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.620955 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.629076 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.629124 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.629138 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.629159 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.629173 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.634081 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.652264 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.677819 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.692838 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.703953 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.704485 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.704546 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.704565 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc772\" (UniqueName: \"kubernetes.io/projected/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-kube-api-access-fc772\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.704583 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.716768 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.732828 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.732883 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.732895 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.732916 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.732931 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.732902 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.760978 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.775926 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.786677 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.805164 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.805263 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.805302 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.805323 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc772\" (UniqueName: \"kubernetes.io/projected/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-kube-api-access-fc772\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.806403 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.806463 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.807663 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.811861 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.825774 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc772\" (UniqueName: \"kubernetes.io/projected/47bf1eb8-79e2-46d7-b4b0-1b0947872fa0-kube-api-access-fc772\") pod \"ovnkube-control-plane-749d76644c-ft6s9\" (UID: \"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.826204 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.841183 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:28Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.875996 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.876045 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.876058 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.876075 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.876089 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.901057 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.979961 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.980005 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.980018 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.980034 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:28 crc kubenswrapper[4992]: I1004 04:33:28.980044 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:28Z","lastTransitionTime":"2025-10-04T04:33:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.082021 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.082058 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.082066 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.082081 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.082093 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.184647 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.184698 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.184709 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.184726 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.184737 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.287745 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.287778 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.287790 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.287808 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.287821 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.318454 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.318497 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:29 crc kubenswrapper[4992]: E1004 04:33:29.318624 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:29 crc kubenswrapper[4992]: E1004 04:33:29.318731 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.390093 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.390127 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.390137 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.390150 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.390159 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.493403 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.493428 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.493436 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.493449 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.493459 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.586771 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/0.log" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.591857 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f" exitCode=1 Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.591880 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.593738 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" event={"ID":"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0","Type":"ContainerStarted","Data":"292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.593773 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" event={"ID":"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0","Type":"ContainerStarted","Data":"b6b63c4031a0fe6b46bca5581ae4415be8424bc62b464f15eed6a699adb6cb93"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.594466 4992 scope.go:117] "RemoveContainer" containerID="e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.594961 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.594998 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.595007 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.595021 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.595029 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.606730 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.619252 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.631176 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.648587 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.668885 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.684257 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.693335 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-6jlms"] Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.694056 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:29 crc kubenswrapper[4992]: E1004 04:33:29.694145 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.697769 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.697829 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.697841 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.697856 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.697869 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.709332 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.724391 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.737264 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.766166 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"kPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 04:33:29.167255 6299 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:33:29.167276 6299 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:33:29.167288 6299 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 04:33:29.167301 6299 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 04:33:29.167298 6299 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167428 6299 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167318 6299 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:33:29.168270 6299 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 04:33:29.168307 6299 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 04:33:29.168326 6299 factory.go:656] Stopping watch factory\\\\nI1004 04:33:29.168342 6299 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:33:29.168337 6299 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.781434 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.791662 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.800675 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.800722 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.800733 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.800752 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.800765 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.806902 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.819264 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhsvj\" (UniqueName: \"kubernetes.io/projected/2ef62768-7ed6-4ce3-9d34-6e6743681405-kube-api-access-nhsvj\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.819307 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.822614 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.835677 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.854924 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.867323 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.878462 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.889037 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.898587 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.902921 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.902948 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.902977 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.902992 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.903001 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:29Z","lastTransitionTime":"2025-10-04T04:33:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.908772 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.919791 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.919991 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhsvj\" (UniqueName: \"kubernetes.io/projected/2ef62768-7ed6-4ce3-9d34-6e6743681405-kube-api-access-nhsvj\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.920046 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:29 crc kubenswrapper[4992]: E1004 04:33:29.920193 4992 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:29 crc kubenswrapper[4992]: E1004 04:33:29.920262 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs podName:2ef62768-7ed6-4ce3-9d34-6e6743681405 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:30.420244427 +0000 UTC m=+44.267919895 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs") pod "network-metrics-daemon-6jlms" (UID: "2ef62768-7ed6-4ce3-9d34-6e6743681405") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.934179 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.937974 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhsvj\" (UniqueName: \"kubernetes.io/projected/2ef62768-7ed6-4ce3-9d34-6e6743681405-kube-api-access-nhsvj\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.946624 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.956900 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.970490 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.988266 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:29 crc kubenswrapper[4992]: I1004 04:33:29.999679 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:29Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.005839 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.005877 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.005886 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.005915 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.005925 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.008894 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.020890 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.034969 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.052922 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"kPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 04:33:29.167255 6299 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:33:29.167276 6299 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:33:29.167288 6299 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 04:33:29.167301 6299 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 04:33:29.167298 6299 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167428 6299 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167318 6299 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:33:29.168270 6299 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 04:33:29.168307 6299 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 04:33:29.168326 6299 factory.go:656] Stopping watch factory\\\\nI1004 04:33:29.168342 6299 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:33:29.168337 6299 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.064455 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.108563 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.108594 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.108604 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.108619 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.108629 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.211524 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.211561 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.211571 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.211586 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.211597 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.313969 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.314023 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.314033 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.314052 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.314067 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.318595 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:30 crc kubenswrapper[4992]: E1004 04:33:30.318738 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.417506 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.417558 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.417576 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.417603 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.417619 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.424966 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:30 crc kubenswrapper[4992]: E1004 04:33:30.425191 4992 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:30 crc kubenswrapper[4992]: E1004 04:33:30.425331 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs podName:2ef62768-7ed6-4ce3-9d34-6e6743681405 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:31.425294869 +0000 UTC m=+45.272970417 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs") pod "network-metrics-daemon-6jlms" (UID: "2ef62768-7ed6-4ce3-9d34-6e6743681405") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.520505 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.520572 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.520583 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.520621 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.520638 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.599723 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" event={"ID":"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0","Type":"ContainerStarted","Data":"07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.602652 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/0.log" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.611764 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.612050 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.622702 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.623608 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.623643 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.623653 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.623672 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.623681 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.634835 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.647346 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.661132 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.682376 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"kPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 04:33:29.167255 6299 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:33:29.167276 6299 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:33:29.167288 6299 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 04:33:29.167301 6299 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 04:33:29.167298 6299 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167428 6299 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167318 6299 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:33:29.168270 6299 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 04:33:29.168307 6299 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 04:33:29.168326 6299 factory.go:656] Stopping watch factory\\\\nI1004 04:33:29.168342 6299 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:33:29.168337 6299 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.697869 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.712389 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.727224 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.727528 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.727621 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.727702 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.727778 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.727702 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.744708 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.761187 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.773600 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.794966 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.809953 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.825604 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.829708 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.829749 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.829758 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.829773 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.829785 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.840089 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.851412 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.864959 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.878517 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.890684 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.902526 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.913238 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.927946 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.932290 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.932326 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.932337 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.932352 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.932382 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:30Z","lastTransitionTime":"2025-10-04T04:33:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.939997 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.954266 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.965579 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:30 crc kubenswrapper[4992]: I1004 04:33:30.980106 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:30Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.029764 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.029883 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:33:47.029860953 +0000 UTC m=+60.877536421 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.034278 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.034326 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.034337 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.034358 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.034384 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.034470 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.048695 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.065656 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.088936 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.106745 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"kPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 04:33:29.167255 6299 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:33:29.167276 6299 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:33:29.167288 6299 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 04:33:29.167301 6299 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 04:33:29.167298 6299 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167428 6299 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167318 6299 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:33:29.168270 6299 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 04:33:29.168307 6299 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 04:33:29.168326 6299 factory.go:656] Stopping watch factory\\\\nI1004 04:33:29.168342 6299 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:33:29.168337 6299 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.120942 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.130586 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.130667 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.130699 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.130725 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130811 4992 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130837 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130866 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130874 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:47.130857473 +0000 UTC m=+60.978532941 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130880 4992 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130886 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130905 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130916 4992 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130954 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:47.130945575 +0000 UTC m=+60.978621043 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.130971 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:47.130963585 +0000 UTC m=+60.978639053 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.131003 4992 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.131053 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:47.131035587 +0000 UTC m=+60.978711105 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.132916 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.136248 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.136280 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.136291 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.136307 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.136319 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.157139 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.239591 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.239635 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.239646 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.239662 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.239673 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.319228 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.319229 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.319439 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.319532 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.319228 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.319705 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.342978 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.343012 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.343024 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.343039 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.343050 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.434123 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.434342 4992 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.434477 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs podName:2ef62768-7ed6-4ce3-9d34-6e6743681405 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:33.434454998 +0000 UTC m=+47.282130486 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs") pod "network-metrics-daemon-6jlms" (UID: "2ef62768-7ed6-4ce3-9d34-6e6743681405") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.444899 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.444950 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.444962 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.444981 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.444994 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.547712 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.547773 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.547785 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.547809 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.547822 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.616930 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/1.log" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.617584 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/0.log" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.619875 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee" exitCode=1 Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.619947 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.620021 4992 scope.go:117] "RemoveContainer" containerID="e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.621206 4992 scope.go:117] "RemoveContainer" containerID="91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee" Oct 04 04:33:31 crc kubenswrapper[4992]: E1004 04:33:31.621434 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.639716 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.650338 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.650404 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.650417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.650438 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.650454 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.660609 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"kPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 04:33:29.167255 6299 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:33:29.167276 6299 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:33:29.167288 6299 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 04:33:29.167301 6299 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 04:33:29.167298 6299 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167428 6299 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167318 6299 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:33:29.168270 6299 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 04:33:29.168307 6299 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 04:33:29.168326 6299 factory.go:656] Stopping watch factory\\\\nI1004 04:33:29.168342 6299 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:33:29.168337 6299 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"message\\\":\\\"UID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:30.613228 6512 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:33:30.612423 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.679853 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.691296 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.706668 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.725527 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.741466 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.754469 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.754528 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.754540 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.754562 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.754574 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.756993 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.772611 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.784697 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.799243 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.811628 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.827349 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.850680 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.858106 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.858171 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.858189 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.858209 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.858222 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.868091 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.882841 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.896036 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:31Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.960855 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.960914 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.960929 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.960952 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:31 crc kubenswrapper[4992]: I1004 04:33:31.960965 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:31Z","lastTransitionTime":"2025-10-04T04:33:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.063839 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.063893 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.063908 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.063929 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.063943 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.168521 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.168568 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.168579 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.168599 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.168612 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.272977 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.273486 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.273503 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.273525 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.273540 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.318685 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:32 crc kubenswrapper[4992]: E1004 04:33:32.318956 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.376185 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.376222 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.376232 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.376248 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.376257 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.479214 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.479262 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.479272 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.479288 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.479299 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.582756 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.582828 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.582848 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.582877 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.582894 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.626051 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/1.log" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.685750 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.685801 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.685814 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.685834 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.685848 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.788718 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.788779 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.788791 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.788810 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.788821 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.890603 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.890680 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.890691 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.890710 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.890724 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.979436 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.992144 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:32Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.993414 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.993459 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.993471 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.993487 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:32 crc kubenswrapper[4992]: I1004 04:33:32.993500 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:32Z","lastTransitionTime":"2025-10-04T04:33:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.006391 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.019259 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.037084 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.057785 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.072898 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.084024 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.097181 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.097221 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.097232 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.097248 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.097261 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.099197 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.117849 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.138868 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"kPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 04:33:29.167255 6299 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:33:29.167276 6299 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:33:29.167288 6299 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 04:33:29.167301 6299 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 04:33:29.167298 6299 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167428 6299 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167318 6299 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:33:29.168270 6299 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 04:33:29.168307 6299 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 04:33:29.168326 6299 factory.go:656] Stopping watch factory\\\\nI1004 04:33:29.168342 6299 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:33:29.168337 6299 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"message\\\":\\\"UID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:30.613228 6512 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:33:30.612423 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.154060 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.171074 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.187348 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.200026 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.200078 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.200091 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.200107 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.200120 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.201153 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.211199 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.220139 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.231710 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:33Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.303670 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.303734 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.303747 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.303765 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.303778 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.318327 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:33 crc kubenswrapper[4992]: E1004 04:33:33.318589 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.318668 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.318730 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:33 crc kubenswrapper[4992]: E1004 04:33:33.318747 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:33 crc kubenswrapper[4992]: E1004 04:33:33.318963 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.406932 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.406987 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.406997 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.407015 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.407028 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.459207 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:33 crc kubenswrapper[4992]: E1004 04:33:33.459426 4992 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:33 crc kubenswrapper[4992]: E1004 04:33:33.459908 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs podName:2ef62768-7ed6-4ce3-9d34-6e6743681405 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:37.459877894 +0000 UTC m=+51.307553402 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs") pod "network-metrics-daemon-6jlms" (UID: "2ef62768-7ed6-4ce3-9d34-6e6743681405") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.510069 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.510410 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.510484 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.510585 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.510681 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.613037 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.613077 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.613099 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.613114 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.613123 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.715547 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.715590 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.715600 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.715616 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.715625 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.818092 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.818134 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.818145 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.818162 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.818172 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.920613 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.920655 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.920666 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.920697 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:33 crc kubenswrapper[4992]: I1004 04:33:33.920707 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:33Z","lastTransitionTime":"2025-10-04T04:33:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.023101 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.023148 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.023157 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.023172 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.023188 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.125769 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.125854 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.125872 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.125898 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.125915 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.228231 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.228290 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.228304 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.228325 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.228339 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.319100 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:34 crc kubenswrapper[4992]: E1004 04:33:34.319207 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.330398 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.330444 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.330455 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.330473 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.330483 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.335037 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.352511 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.365171 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.379179 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.401290 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.414967 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.428877 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.432752 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.432806 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.432816 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.432832 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.432842 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.444455 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.457922 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.478095 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.491761 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.502562 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.518671 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.534952 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.536665 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.536707 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.536721 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.536741 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.536755 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.555042 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e55bf2c813658212f8a85ebd0d3ef1475a2389bacf1f4daf721447abf378937f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"kPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 04:33:29.167255 6299 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 04:33:29.167276 6299 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1004 04:33:29.167288 6299 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 04:33:29.167301 6299 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1004 04:33:29.167298 6299 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167428 6299 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:33:29.167318 6299 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 04:33:29.168270 6299 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1004 04:33:29.168307 6299 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 04:33:29.168326 6299 factory.go:656] Stopping watch factory\\\\nI1004 04:33:29.168342 6299 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:33:29.168337 6299 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"message\\\":\\\"UID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:30.613228 6512 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:33:30.612423 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.569449 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.581635 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:34Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.639429 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.639496 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.639507 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.639527 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.639542 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.742174 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.742235 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.742246 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.742262 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.742272 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.844770 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.844823 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.844833 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.844847 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.844860 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.946864 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.946898 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.946907 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.946920 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:34 crc kubenswrapper[4992]: I1004 04:33:34.946936 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:34Z","lastTransitionTime":"2025-10-04T04:33:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.049595 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.049630 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.049639 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.049654 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.049663 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.153017 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.153086 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.153108 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.153140 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.153165 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.256280 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.256355 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.256415 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.256443 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.256466 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.260564 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.261522 4992 scope.go:117] "RemoveContainer" containerID="91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee" Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.261760 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.277496 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.296320 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.312723 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.318568 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.318685 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.318780 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.318705 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.318940 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.319052 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.326640 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.341141 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.354024 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.359399 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.359485 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.359511 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.359549 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.359571 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.372402 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.386891 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.400327 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.413051 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.430104 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.454961 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.462552 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.462624 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.462636 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.462661 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.462676 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.468962 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.484879 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.501451 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.520730 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"message\\\":\\\"UID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:30.613228 6512 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:33:30.612423 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.534007 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.534050 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.534058 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.534074 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.534085 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.544584 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.546081 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.549294 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.549318 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.549326 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.549338 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.549348 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.561168 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.564951 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.564981 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.564989 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.565004 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.565015 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.576831 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.580951 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.580985 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.580995 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.581012 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.581025 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.593034 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.597346 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.597424 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.597440 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.597459 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.597472 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.609219 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:35Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:35 crc kubenswrapper[4992]: E1004 04:33:35.609342 4992 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.611195 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.611256 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.611264 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.611282 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.611295 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.713789 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.713838 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.713851 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.713866 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.713877 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.816430 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.816474 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.816488 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.816506 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.816518 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.918550 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.918590 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.918599 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.918612 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:35 crc kubenswrapper[4992]: I1004 04:33:35.918621 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:35Z","lastTransitionTime":"2025-10-04T04:33:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.020855 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.020885 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.020894 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.020908 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.020916 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.123411 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.123475 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.123493 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.123517 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.123534 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.226092 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.226166 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.226180 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.226201 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.226215 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.319049 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:36 crc kubenswrapper[4992]: E1004 04:33:36.319277 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.328254 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.328309 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.328320 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.328343 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.328372 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.431523 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.431582 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.431604 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.431627 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.431640 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.534071 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.534124 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.534138 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.534160 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.534177 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.638027 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.638090 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.638109 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.638141 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.638164 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.740854 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.740939 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.740956 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.740980 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.740995 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.843957 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.844029 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.844047 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.844064 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.844075 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.947394 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.947479 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.947504 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.947538 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:36 crc kubenswrapper[4992]: I1004 04:33:36.947563 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:36Z","lastTransitionTime":"2025-10-04T04:33:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.055280 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.055423 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.055475 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.056159 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.056865 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.160245 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.160308 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.160326 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.160352 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.160404 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.263511 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.263578 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.263591 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.263607 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.263620 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.318869 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.318961 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.318889 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:37 crc kubenswrapper[4992]: E1004 04:33:37.319169 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:37 crc kubenswrapper[4992]: E1004 04:33:37.319299 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:37 crc kubenswrapper[4992]: E1004 04:33:37.319435 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.366443 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.366488 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.366501 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.366519 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.366534 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.469307 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.469339 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.469348 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.469376 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.469385 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.502275 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:37 crc kubenswrapper[4992]: E1004 04:33:37.502517 4992 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:37 crc kubenswrapper[4992]: E1004 04:33:37.502619 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs podName:2ef62768-7ed6-4ce3-9d34-6e6743681405 nodeName:}" failed. No retries permitted until 2025-10-04 04:33:45.502599503 +0000 UTC m=+59.350274971 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs") pod "network-metrics-daemon-6jlms" (UID: "2ef62768-7ed6-4ce3-9d34-6e6743681405") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.572991 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.573077 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.573094 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.573120 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.573137 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.676323 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.676430 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.676448 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.676501 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.676517 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.779925 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.779979 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.780004 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.780029 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.780049 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.882975 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.883008 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.883017 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.883031 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.883043 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.986124 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.986180 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.986197 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.986221 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:37 crc kubenswrapper[4992]: I1004 04:33:37.986240 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:37Z","lastTransitionTime":"2025-10-04T04:33:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.088931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.088998 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.089015 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.089042 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.089061 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.192000 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.192056 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.192075 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.192095 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.192109 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.295001 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.295063 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.295082 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.295110 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.295131 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.319025 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:38 crc kubenswrapper[4992]: E1004 04:33:38.319215 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.397470 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.397513 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.397525 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.397543 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.397558 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.500504 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.500544 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.500555 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.500569 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.500578 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.603599 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.603649 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.603660 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.603688 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.603703 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.706680 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.706730 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.706776 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.706798 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.706813 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.809340 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.809468 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.809496 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.809527 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.809547 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.912723 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.912808 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.912838 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.912872 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:38 crc kubenswrapper[4992]: I1004 04:33:38.912895 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:38Z","lastTransitionTime":"2025-10-04T04:33:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.015196 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.015278 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.015305 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.015337 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.015409 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.118495 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.119038 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.119077 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.119095 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.119106 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.221114 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.221190 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.221212 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.221239 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.221257 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.318527 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.318639 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.318731 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:39 crc kubenswrapper[4992]: E1004 04:33:39.318721 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:39 crc kubenswrapper[4992]: E1004 04:33:39.318932 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:39 crc kubenswrapper[4992]: E1004 04:33:39.319116 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.324177 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.324223 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.324239 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.324259 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.324276 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.428282 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.428342 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.428360 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.428393 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.428413 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.530800 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.530840 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.530852 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.530867 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.530878 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.633631 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.633683 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.633696 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.633716 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.633730 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.736029 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.736073 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.736087 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.736105 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.736117 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.838490 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.838538 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.838552 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.838571 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.838584 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.940649 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.940686 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.940697 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.940717 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:39 crc kubenswrapper[4992]: I1004 04:33:39.940743 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:39Z","lastTransitionTime":"2025-10-04T04:33:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.043118 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.043172 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.043189 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.043210 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.043224 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.146298 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.146330 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.146345 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.146378 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.146390 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.249114 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.249162 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.249171 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.249188 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.249198 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.319167 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:40 crc kubenswrapper[4992]: E1004 04:33:40.319342 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.352057 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.352109 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.352124 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.352143 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.352157 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.455198 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.455239 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.455248 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.455264 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.455275 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.557319 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.557378 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.557392 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.557409 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.557422 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.659561 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.659606 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.659614 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.659632 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.659643 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.762466 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.762502 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.762511 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.762541 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.762554 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.865543 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.865579 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.865591 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.865611 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.865623 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.967751 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.967789 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.967801 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.967820 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:40 crc kubenswrapper[4992]: I1004 04:33:40.967835 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:40Z","lastTransitionTime":"2025-10-04T04:33:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.070059 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.070098 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.070109 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.070129 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.070142 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.172929 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.172975 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.172993 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.173014 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.173028 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.276596 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.276666 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.276691 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.276720 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.276741 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.318631 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.318707 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.318766 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:41 crc kubenswrapper[4992]: E1004 04:33:41.318924 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:41 crc kubenswrapper[4992]: E1004 04:33:41.319173 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:41 crc kubenswrapper[4992]: E1004 04:33:41.319511 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.379607 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.379651 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.379661 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.379677 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.379686 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.482206 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.482246 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.482258 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.482276 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.482292 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.584630 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.584676 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.584687 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.584704 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.584715 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.686683 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.686729 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.686741 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.686759 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.686772 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.789660 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.789691 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.789699 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.789740 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.789760 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.892504 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.892547 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.892559 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.892576 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.892588 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.995140 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.995183 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.995194 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.995209 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:41 crc kubenswrapper[4992]: I1004 04:33:41.995221 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:41Z","lastTransitionTime":"2025-10-04T04:33:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.097896 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.097954 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.097971 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.097995 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.098010 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.200934 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.200990 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.201000 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.201017 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.201029 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.303640 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.303680 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.303693 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.303710 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.303722 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.319343 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:42 crc kubenswrapper[4992]: E1004 04:33:42.319577 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.406644 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.406697 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.406714 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.406737 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.406753 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.509846 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.509883 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.509894 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.509913 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.509925 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.612418 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.612458 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.612468 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.612483 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.612497 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.715015 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.715082 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.715105 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.715135 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.715159 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.790351 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.802200 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.809087 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.818218 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.818414 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.818449 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.818482 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.818507 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.827599 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.843917 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.855044 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.874162 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.890454 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.902710 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.914140 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.920144 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.920174 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.920182 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.920197 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.920206 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:42Z","lastTransitionTime":"2025-10-04T04:33:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.929379 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.945515 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.959574 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.973142 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:42 crc kubenswrapper[4992]: I1004 04:33:42.990692 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:42Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.010597 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.022462 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.022687 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.022798 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.022915 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.023033 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.030706 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"message\\\":\\\"UID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:30.613228 6512 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:33:30.612423 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.047644 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.061218 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:43Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.125638 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.125686 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.125695 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.125712 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.125725 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.228774 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.228814 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.228832 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.228852 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.228864 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.318354 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.318413 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:43 crc kubenswrapper[4992]: E1004 04:33:43.318906 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.318458 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:43 crc kubenswrapper[4992]: E1004 04:33:43.318975 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:43 crc kubenswrapper[4992]: E1004 04:33:43.318844 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.332162 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.332203 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.332216 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.332233 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.332246 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.435677 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.435717 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.435732 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.435755 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.435772 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.538864 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.538909 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.538921 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.538937 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.538952 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.641729 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.641785 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.641806 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.641837 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.641859 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.744637 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.744677 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.744688 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.744704 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.744714 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.847838 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.847890 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.847902 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.847920 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.847933 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.951437 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.951491 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.951503 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.951519 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:43 crc kubenswrapper[4992]: I1004 04:33:43.951531 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:43Z","lastTransitionTime":"2025-10-04T04:33:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.054461 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.054531 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.054543 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.054571 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.054585 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.157226 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.157259 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.157268 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.157282 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.157290 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.259802 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.259868 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.259895 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.259925 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.259947 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.319829 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:44 crc kubenswrapper[4992]: E1004 04:33:44.320045 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.340125 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.355583 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.362521 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.362556 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.362566 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.362580 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.362590 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.368954 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.381783 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.392776 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.403612 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.425269 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.439226 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.452832 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.464006 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.464043 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.464057 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.464075 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.464087 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.472682 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.481879 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.492223 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.504472 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.516405 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.526533 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.537220 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.576811 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.583605 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.583650 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.583660 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.583678 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.583692 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.606984 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"message\\\":\\\"UID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:30.613228 6512 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:33:30.612423 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:44Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.685517 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.685569 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.685581 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.685599 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.685611 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.788240 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.788273 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.788306 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.788321 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.788330 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.891001 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.891030 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.891040 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.891053 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.891062 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.993658 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.993695 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.993705 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.993719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:44 crc kubenswrapper[4992]: I1004 04:33:44.993729 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:44Z","lastTransitionTime":"2025-10-04T04:33:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.095548 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.095597 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.095610 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.095628 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.095642 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.197525 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.197554 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.197563 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.197576 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.197585 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.300502 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.300544 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.300556 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.300572 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.300584 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.318424 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.318490 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.318568 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.318723 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.318497 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.318919 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.404187 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.404249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.404263 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.404282 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.404296 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.507036 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.507098 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.507115 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.507136 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.507153 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.588486 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.588635 4992 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.588697 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs podName:2ef62768-7ed6-4ce3-9d34-6e6743681405 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:01.588674887 +0000 UTC m=+75.436350375 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs") pod "network-metrics-daemon-6jlms" (UID: "2ef62768-7ed6-4ce3-9d34-6e6743681405") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.610542 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.610596 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.610609 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.610625 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.610640 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.713783 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.713833 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.713849 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.713873 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.713890 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.813256 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.813311 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.813333 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.813353 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.813385 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.826445 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.829711 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.829744 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.829754 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.829772 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.829783 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.840786 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.845297 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.845343 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.845375 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.845401 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.845417 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.859531 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.862328 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.862382 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.862400 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.862416 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.862428 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.872935 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.876262 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.876299 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.876311 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.876330 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.876342 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.887052 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:45Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:45 crc kubenswrapper[4992]: E1004 04:33:45.887198 4992 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.888808 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.888848 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.888861 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.888879 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.888890 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.991661 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.991709 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.991718 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.991734 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:45 crc kubenswrapper[4992]: I1004 04:33:45.991744 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:45Z","lastTransitionTime":"2025-10-04T04:33:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.095010 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.095063 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.095076 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.095095 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.095110 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.198040 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.199249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.199429 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.199577 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.199697 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.302671 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.303022 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.303104 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.303169 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.303230 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.319266 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:46 crc kubenswrapper[4992]: E1004 04:33:46.319827 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.320121 4992 scope.go:117] "RemoveContainer" containerID="91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.405834 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.406086 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.406096 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.406110 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.406119 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.508569 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.508604 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.508615 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.508632 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.508645 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.611859 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.611940 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.612010 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.612045 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.612065 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.680506 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/1.log" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.683239 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.684424 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.696459 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.711545 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.714417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.714491 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.714520 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.714555 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.714578 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.732939 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.753864 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"message\\\":\\\"UID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:30.613228 6512 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:33:30.612423 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.773270 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.786567 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.803148 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.817471 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.817766 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.817884 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.817992 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.818096 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.825514 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.837730 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.848737 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.860767 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.873996 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.888825 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.904335 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.922186 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.922227 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.922239 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.922257 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.922269 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:46Z","lastTransitionTime":"2025-10-04T04:33:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.923266 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.951244 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.971861 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:46 crc kubenswrapper[4992]: I1004 04:33:46.990229 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:46Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.029225 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.029257 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.029266 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.029279 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.029289 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.104353 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.104592 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:34:19.10457582 +0000 UTC m=+92.952251288 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.131814 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.132109 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.132172 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.132300 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.132424 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.205832 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.206142 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.207420 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.207442 4992 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.207526 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:19.207499268 +0000 UTC m=+93.055174736 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.207381 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.208036 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.208130 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.208213 4992 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.208248 4992 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.208287 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:19.208279998 +0000 UTC m=+93.055955466 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.208326 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:19.208295148 +0000 UTC m=+93.055970616 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.209130 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.209231 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.209310 4992 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.209456 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:19.209439727 +0000 UTC m=+93.057115195 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.234613 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.234895 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.234976 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.235066 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.235150 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.318642 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.318651 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.318864 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.318964 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.319115 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.319321 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.337643 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.337888 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.337971 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.338093 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.338164 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.440264 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.440301 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.440310 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.440325 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.440334 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.542725 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.542772 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.542783 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.542801 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.542814 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.644942 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.645054 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.645073 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.645102 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.645121 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.688541 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/2.log" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.689187 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/1.log" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.692117 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020" exitCode=1 Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.692178 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.692547 4992 scope.go:117] "RemoveContainer" containerID="91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.693296 4992 scope.go:117] "RemoveContainer" containerID="503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020" Oct 04 04:33:47 crc kubenswrapper[4992]: E1004 04:33:47.693995 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.709707 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.727579 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.740138 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.748067 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.748108 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.748118 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.748133 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.748144 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.755385 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.769596 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.781766 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.794044 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.808803 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.834488 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.851879 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.851959 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.851998 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.852012 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.852032 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.852045 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.866129 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.878774 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.893942 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.913882 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91fb49ae52a46c2c47fe5c19c464e99f3d8b320530b1b8dd0c2c66413bae5cee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"message\\\":\\\"UID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:30.613228 6512 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 04:33:30.612423 6512 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.927203 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.941069 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.951999 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.954841 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.954894 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.954906 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.954930 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.954943 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:47Z","lastTransitionTime":"2025-10-04T04:33:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:47 crc kubenswrapper[4992]: I1004 04:33:47.965206 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.058185 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.058236 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.058249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.058265 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.058276 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.159909 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.159961 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.159973 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.159992 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.160004 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.264696 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.264748 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.264760 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.264777 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.264792 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.318225 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:48 crc kubenswrapper[4992]: E1004 04:33:48.318335 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.366633 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.366677 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.366721 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.366761 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.366773 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.469088 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.469134 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.469144 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.469159 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.469170 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.571343 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.571396 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.571405 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.571421 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.571433 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.674074 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.674144 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.674162 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.674179 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.674192 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.698323 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/2.log" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.702943 4992 scope.go:117] "RemoveContainer" containerID="503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020" Oct 04 04:33:48 crc kubenswrapper[4992]: E1004 04:33:48.703123 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.719280 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.731615 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.746613 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.766678 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.777477 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.777547 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.777565 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.777588 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.777605 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.788727 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.802958 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.814827 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.827118 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.839426 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.849899 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.860862 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.871589 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.880103 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.880140 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.880149 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.880163 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.880174 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.885482 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.899152 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.913082 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.926392 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.940666 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.962539 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:48Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.982253 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.982292 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.982301 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.982315 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:48 crc kubenswrapper[4992]: I1004 04:33:48.982325 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:48Z","lastTransitionTime":"2025-10-04T04:33:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.086207 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.086289 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.086299 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.086318 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.086335 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.190022 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.190069 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.190095 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.190120 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.190137 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.292901 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.292941 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.292958 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.292978 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.292989 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.318474 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:49 crc kubenswrapper[4992]: E1004 04:33:49.318610 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.318496 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.318484 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:49 crc kubenswrapper[4992]: E1004 04:33:49.318685 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:49 crc kubenswrapper[4992]: E1004 04:33:49.318765 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.395595 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.395637 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.395647 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.395662 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.395672 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.497255 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.497302 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.497312 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.497330 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.497342 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.600028 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.600073 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.600090 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.600106 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.600117 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.702893 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.702931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.702941 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.702957 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.702967 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.805346 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.805438 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.805454 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.805469 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.805479 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.908043 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.908097 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.908108 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.908122 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:49 crc kubenswrapper[4992]: I1004 04:33:49.908132 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:49Z","lastTransitionTime":"2025-10-04T04:33:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.010760 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.010798 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.010811 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.010828 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.010844 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.113777 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.113831 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.113840 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.113854 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.113863 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.216946 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.216990 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.217002 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.217018 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.217028 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.318415 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:50 crc kubenswrapper[4992]: E1004 04:33:50.318525 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.318632 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.318653 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.318663 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.318679 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.318690 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.420992 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.421036 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.421048 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.421064 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.421076 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.523510 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.523540 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.523548 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.523562 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.523572 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.625228 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.625271 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.625281 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.625297 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.625308 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.728287 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.728323 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.728342 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.728372 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.728386 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.830517 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.830557 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.830568 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.830585 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.830596 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.933824 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.933876 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.933887 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.933907 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:50 crc kubenswrapper[4992]: I1004 04:33:50.933920 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:50Z","lastTransitionTime":"2025-10-04T04:33:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.036280 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.036332 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.036343 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.036391 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.036409 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.139163 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.139197 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.139207 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.139221 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.139231 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.241931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.241979 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.241989 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.242007 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.242017 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.318199 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.318260 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.318199 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:51 crc kubenswrapper[4992]: E1004 04:33:51.319150 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:51 crc kubenswrapper[4992]: E1004 04:33:51.319286 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:51 crc kubenswrapper[4992]: E1004 04:33:51.319542 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.344460 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.344500 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.344511 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.344527 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.344539 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.447685 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.447738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.447750 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.447768 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.447781 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.550398 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.550440 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.550448 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.550461 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.550470 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.652340 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.652398 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.652409 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.652425 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.652437 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.754402 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.754451 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.754461 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.754474 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.754483 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.856306 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.856351 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.856381 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.856399 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.856409 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.958783 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.958820 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.958828 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.958844 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:51 crc kubenswrapper[4992]: I1004 04:33:51.958853 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:51Z","lastTransitionTime":"2025-10-04T04:33:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.066044 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.066124 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.066146 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.066176 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.066197 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.168669 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.168719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.168731 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.168745 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.168754 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.271030 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.271075 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.271083 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.271097 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.271106 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.318862 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:52 crc kubenswrapper[4992]: E1004 04:33:52.319024 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.374056 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.374092 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.374101 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.374117 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.374127 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.476666 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.476733 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.476741 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.476757 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.476766 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.579461 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.579505 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.579513 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.579530 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.579539 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.681979 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.682024 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.682035 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.682053 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.682065 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.783970 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.784013 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.784026 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.784043 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.784053 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.886258 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.886304 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.886317 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.886334 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.886346 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.989136 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.989186 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.989201 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.989222 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:52 crc kubenswrapper[4992]: I1004 04:33:52.989240 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:52Z","lastTransitionTime":"2025-10-04T04:33:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.091872 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.091911 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.091924 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.091944 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.091958 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.194043 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.194092 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.194104 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.194118 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.194128 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.296011 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.296059 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.296070 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.296092 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.296105 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.318385 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.318422 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.318452 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:53 crc kubenswrapper[4992]: E1004 04:33:53.318543 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:53 crc kubenswrapper[4992]: E1004 04:33:53.318685 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:53 crc kubenswrapper[4992]: E1004 04:33:53.318814 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.398328 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.398540 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.398571 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.398600 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.398626 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.501420 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.501462 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.501470 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.501485 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.501494 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.603703 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.603734 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.603748 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.603789 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.603801 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.706287 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.706331 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.706339 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.706354 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.706392 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.809811 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.809978 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.810046 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.810072 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.810914 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.914134 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.914210 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.914222 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.914240 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:53 crc kubenswrapper[4992]: I1004 04:33:53.914253 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:53Z","lastTransitionTime":"2025-10-04T04:33:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.015922 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.015961 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.015972 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.015988 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.015999 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.119141 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.119183 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.119193 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.119208 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.119220 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.221882 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.221949 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.221965 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.221988 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.222011 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.318190 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:54 crc kubenswrapper[4992]: E1004 04:33:54.318329 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.324401 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.324459 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.324473 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.324496 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.324508 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.330994 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.341865 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.350885 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.361522 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.375334 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.395757 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.409678 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.424259 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.427468 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.427495 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.427504 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.427517 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.427526 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.438605 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.452182 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.463551 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.475718 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.497864 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.510408 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.524220 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.530353 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.530410 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.530424 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.530440 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.530450 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.536611 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.544686 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.554307 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:54Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.632195 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.632243 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.632255 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.632271 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.632282 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.735122 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.735180 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.735198 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.735224 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.735242 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.838417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.838931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.838944 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.838964 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.838978 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.941153 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.941212 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.941224 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.941248 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:54 crc kubenswrapper[4992]: I1004 04:33:54.941261 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:54Z","lastTransitionTime":"2025-10-04T04:33:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.043704 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.043729 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.043737 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.043750 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.043759 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.147339 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.147429 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.147474 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.147501 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.147518 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.250458 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.250498 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.250509 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.250528 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.250544 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.318395 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.318476 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:55 crc kubenswrapper[4992]: E1004 04:33:55.318532 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.318400 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:55 crc kubenswrapper[4992]: E1004 04:33:55.318614 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:55 crc kubenswrapper[4992]: E1004 04:33:55.318708 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.353700 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.353756 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.353768 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.353792 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.353803 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.456690 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.456759 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.456774 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.456797 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.456812 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.560428 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.560489 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.560501 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.560517 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.560529 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.662341 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.662393 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.662403 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.662417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.662428 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.766059 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.766136 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.766163 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.766193 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.766219 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.869079 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.869146 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.869159 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.869213 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.869227 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.972678 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.972755 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.972798 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.972819 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:55 crc kubenswrapper[4992]: I1004 04:33:55.972833 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:55Z","lastTransitionTime":"2025-10-04T04:33:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.076028 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.076087 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.076102 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.076126 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.076142 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.180103 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.180159 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.180171 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.180194 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.180209 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.208619 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.208712 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.208730 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.208759 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.208775 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: E1004 04:33:56.224091 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:56Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.229734 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.229785 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.229807 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.229841 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.229860 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: E1004 04:33:56.247808 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:56Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.254174 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.254218 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.254230 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.254250 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.254265 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: E1004 04:33:56.267530 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:56Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.272962 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.273029 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.273111 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.273166 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.273192 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: E1004 04:33:56.286701 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:56Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.291237 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.291282 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.291296 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.291316 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.291327 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: E1004 04:33:56.304842 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:56Z is after 2025-08-24T17:21:41Z" Oct 04 04:33:56 crc kubenswrapper[4992]: E1004 04:33:56.305109 4992 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.306988 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.307040 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.307057 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.307081 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.307095 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.318402 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:56 crc kubenswrapper[4992]: E1004 04:33:56.318542 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.409839 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.409887 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.409897 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.409918 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.409929 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.512300 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.512373 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.512384 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.512405 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.512418 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.615848 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.615893 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.615904 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.615925 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.615942 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.719531 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.719824 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.719844 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.719877 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.719897 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.823483 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.823534 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.823546 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.823564 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.823576 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.926770 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.926839 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.926866 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.926895 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:56 crc kubenswrapper[4992]: I1004 04:33:56.926914 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:56Z","lastTransitionTime":"2025-10-04T04:33:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.030749 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.030804 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.030817 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.030838 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.030852 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.134408 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.134457 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.134468 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.134490 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.134501 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.237518 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.237582 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.237606 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.237636 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.237656 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.318683 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.318695 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:57 crc kubenswrapper[4992]: E1004 04:33:57.318885 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.318714 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:57 crc kubenswrapper[4992]: E1004 04:33:57.319075 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:57 crc kubenswrapper[4992]: E1004 04:33:57.319208 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.341491 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.341564 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.341580 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.341601 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.341616 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.444096 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.444142 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.444154 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.444172 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.444183 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.546411 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.546472 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.546491 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.546518 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.546542 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.649207 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.649249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.649260 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.649280 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.649290 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.751878 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.751912 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.751921 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.751936 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.751946 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.854925 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.854979 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.854992 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.855011 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.855023 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.957684 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.957766 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.957788 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.957812 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:57 crc kubenswrapper[4992]: I1004 04:33:57.957832 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:57Z","lastTransitionTime":"2025-10-04T04:33:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.061063 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.061130 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.061144 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.061167 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.061181 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.164326 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.164403 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.164417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.164435 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.164447 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.266681 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.266736 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.266745 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.266761 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.266772 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.318593 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:33:58 crc kubenswrapper[4992]: E1004 04:33:58.318760 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.369542 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.369586 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.369596 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.369611 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.369623 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.471992 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.472028 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.472038 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.472053 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.472065 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.574070 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.574105 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.574113 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.574128 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.574139 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.676979 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.677020 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.677028 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.677043 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.677053 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.779600 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.779809 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.779868 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.779961 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.780023 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.882558 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.882587 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.882596 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.882610 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.882620 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.985370 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.985418 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.985430 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.985447 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:58 crc kubenswrapper[4992]: I1004 04:33:58.985460 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:58Z","lastTransitionTime":"2025-10-04T04:33:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.087966 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.088006 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.088017 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.088032 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.088041 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.189742 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.189782 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.189793 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.189807 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.189817 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.291752 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.291795 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.291807 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.291823 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.291835 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.318235 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.318293 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:33:59 crc kubenswrapper[4992]: E1004 04:33:59.318351 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.318293 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:33:59 crc kubenswrapper[4992]: E1004 04:33:59.318449 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:33:59 crc kubenswrapper[4992]: E1004 04:33:59.318545 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.394374 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.394417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.394427 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.394441 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.394453 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.496653 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.496697 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.496708 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.496722 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.496731 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.599309 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.599348 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.599380 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.599399 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.599413 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.701555 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.701596 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.701604 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.701620 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.701630 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.804186 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.804255 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.804276 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.804298 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.804314 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.906696 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.906743 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.906755 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.906775 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:33:59 crc kubenswrapper[4992]: I1004 04:33:59.906787 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:33:59Z","lastTransitionTime":"2025-10-04T04:33:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.009068 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.009116 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.009127 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.009145 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.009158 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.111582 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.111618 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.111627 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.111641 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.111653 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.213974 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.214015 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.214024 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.214040 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.214055 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.316324 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.316385 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.316396 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.316410 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.316421 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.319273 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:00 crc kubenswrapper[4992]: E1004 04:34:00.319506 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.418093 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.418132 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.418144 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.418162 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.418172 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.520093 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.520135 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.520143 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.520158 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.520167 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.623058 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.623098 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.623109 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.623126 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.623141 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.725203 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.725243 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.725251 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.725264 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.725274 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.828329 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.828431 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.828443 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.828461 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.828473 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.930760 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.930807 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.930817 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.930832 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:00 crc kubenswrapper[4992]: I1004 04:34:00.930844 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:00Z","lastTransitionTime":"2025-10-04T04:34:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.034157 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.034214 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.034230 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.034256 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.034270 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.136813 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.136841 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.136851 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.136866 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.136876 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.238869 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.238906 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.238918 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.238934 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.238945 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.318763 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.318827 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:01 crc kubenswrapper[4992]: E1004 04:34:01.318880 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:01 crc kubenswrapper[4992]: E1004 04:34:01.318920 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.318828 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:01 crc kubenswrapper[4992]: E1004 04:34:01.319077 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.341141 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.341193 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.341204 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.341240 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.341253 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.443549 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.443590 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.443600 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.443616 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.443625 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.546588 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.546651 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.546679 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.546702 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.546720 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.649402 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.649452 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.649465 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.649484 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.649497 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.656901 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:01 crc kubenswrapper[4992]: E1004 04:34:01.657082 4992 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:01 crc kubenswrapper[4992]: E1004 04:34:01.657186 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs podName:2ef62768-7ed6-4ce3-9d34-6e6743681405 nodeName:}" failed. No retries permitted until 2025-10-04 04:34:33.657162335 +0000 UTC m=+107.504837893 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs") pod "network-metrics-daemon-6jlms" (UID: "2ef62768-7ed6-4ce3-9d34-6e6743681405") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.751653 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.751700 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.751715 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.751738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.751754 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.854929 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.854966 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.854975 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.854995 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.855003 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.957471 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.957504 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.957513 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.957526 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:01 crc kubenswrapper[4992]: I1004 04:34:01.957535 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:01Z","lastTransitionTime":"2025-10-04T04:34:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.059696 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.059726 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.059735 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.059748 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.059758 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.162401 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.162439 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.162453 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.162470 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.162483 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.266062 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.266197 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.266223 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.266249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.266269 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.319112 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:02 crc kubenswrapper[4992]: E1004 04:34:02.319246 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.368832 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.368874 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.368883 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.368897 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.368908 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.471208 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.471251 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.471262 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.471278 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.471287 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.573126 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.573171 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.573234 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.573256 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.573270 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.675463 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.675510 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.675567 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.675584 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.675598 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.777679 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.777712 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.777722 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.777737 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.777749 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.880582 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.880639 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.880658 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.880681 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.880697 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.982996 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.983049 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.983062 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.983078 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:02 crc kubenswrapper[4992]: I1004 04:34:02.983089 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:02Z","lastTransitionTime":"2025-10-04T04:34:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.085178 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.085225 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.085233 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.085249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.085259 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.187513 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.187561 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.187571 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.187590 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.187602 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.289941 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.289972 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.289981 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.289997 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.290005 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.318321 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.318384 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.318419 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:03 crc kubenswrapper[4992]: E1004 04:34:03.318483 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:03 crc kubenswrapper[4992]: E1004 04:34:03.318783 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:03 crc kubenswrapper[4992]: E1004 04:34:03.318851 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.319215 4992 scope.go:117] "RemoveContainer" containerID="503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020" Oct 04 04:34:03 crc kubenswrapper[4992]: E1004 04:34:03.319502 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.393028 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.393074 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.393093 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.393116 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.393133 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.495461 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.495506 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.495515 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.495531 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.495541 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.597903 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.597946 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.597959 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.597976 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.597989 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.700898 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.700949 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.700962 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.700980 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.700995 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.803921 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.803977 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.803987 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.804008 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.804020 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.906630 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.906666 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.906675 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.906688 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:03 crc kubenswrapper[4992]: I1004 04:34:03.906697 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:03Z","lastTransitionTime":"2025-10-04T04:34:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.008580 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.008616 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.008628 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.008643 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.008655 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.111447 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.111487 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.111498 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.111514 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.111525 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.214612 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.214667 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.214690 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.214711 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.214722 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.316857 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.316886 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.316894 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.316908 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.316917 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.318491 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:04 crc kubenswrapper[4992]: E1004 04:34:04.318620 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.333185 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.345033 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.359057 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.377015 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.395304 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.409223 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.420425 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.420542 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.420555 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.420571 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.420583 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.425216 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.443340 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.457098 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.469213 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.481173 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.489919 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.499957 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.513449 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.523047 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.523086 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.523099 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.523115 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.523141 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.525552 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.535556 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.546101 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.563421 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.626981 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.627022 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.627031 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.627047 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.627058 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.729925 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.729973 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.729986 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.730005 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.730018 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.753767 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/0.log" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.753836 4992 generic.go:334] "Generic (PLEG): container finished" podID="2b6879c5-46f4-4d87-959b-03dff08cff4c" containerID="7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79" exitCode=1 Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.753868 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhq64" event={"ID":"2b6879c5-46f4-4d87-959b-03dff08cff4c","Type":"ContainerDied","Data":"7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.754403 4992 scope.go:117] "RemoveContainer" containerID="7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.767605 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.777548 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.792935 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:03Z\\\",\\\"message\\\":\\\"2025-10-04T04:33:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e\\\\n2025-10-04T04:33:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e to /host/opt/cni/bin/\\\\n2025-10-04T04:33:18Z [verbose] multus-daemon started\\\\n2025-10-04T04:33:18Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:34:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.817949 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.833074 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.833120 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.833134 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.833164 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.833177 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.836974 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.851542 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.863786 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.880945 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.903542 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.919048 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.931144 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.935764 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.935800 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.935809 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.935823 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.935836 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:04Z","lastTransitionTime":"2025-10-04T04:34:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.941029 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.956008 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.969466 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.980718 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:04 crc kubenswrapper[4992]: I1004 04:34:04.992435 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:04Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.005272 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.018067 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.038806 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.038867 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.038878 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.038903 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.038922 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.141352 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.141442 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.141455 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.141479 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.141491 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.244107 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.244180 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.244228 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.244254 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.244268 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.318510 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.318565 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.318531 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:05 crc kubenswrapper[4992]: E1004 04:34:05.318662 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:05 crc kubenswrapper[4992]: E1004 04:34:05.318735 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:05 crc kubenswrapper[4992]: E1004 04:34:05.318827 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.348014 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.348081 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.348094 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.348113 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.348126 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.450827 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.450868 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.450901 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.450922 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.450933 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.553711 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.553753 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.553765 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.553785 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.553798 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.656618 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.656682 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.656697 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.656717 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.656728 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.758956 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.758999 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.759012 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.759032 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.759046 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.759603 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/0.log" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.759685 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhq64" event={"ID":"2b6879c5-46f4-4d87-959b-03dff08cff4c","Type":"ContainerStarted","Data":"e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.774242 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.788731 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.804425 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.821242 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.835900 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.846229 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.862265 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.862312 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.862324 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.862339 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.862348 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.866746 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.880847 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.891971 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.905349 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.914996 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.925982 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:03Z\\\",\\\"message\\\":\\\"2025-10-04T04:33:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e\\\\n2025-10-04T04:33:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e to /host/opt/cni/bin/\\\\n2025-10-04T04:33:18Z [verbose] multus-daemon started\\\\n2025-10-04T04:33:18Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:34:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.937322 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.948973 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.959249 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.964460 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.964494 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.964504 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.964521 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.964533 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:05Z","lastTransitionTime":"2025-10-04T04:34:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.973835 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:05 crc kubenswrapper[4992]: I1004 04:34:05.988115 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:05Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.008149 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.067075 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.067100 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.067109 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.067123 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.067132 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.169679 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.169719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.169728 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.169742 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.169752 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.272957 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.273002 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.273018 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.273041 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.273058 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.318934 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:06 crc kubenswrapper[4992]: E1004 04:34:06.319102 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.375044 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.375074 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.375084 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.375098 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.375109 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.426949 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.427010 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.427027 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.427052 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.427068 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: E1004 04:34:06.439846 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.443716 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.443757 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.443769 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.443786 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.443799 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: E1004 04:34:06.456729 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.460017 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.460062 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.460074 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.460093 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.460106 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: E1004 04:34:06.470496 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.473605 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.473660 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.473675 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.473697 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.473712 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: E1004 04:34:06.484410 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.488038 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.488076 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.488085 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.488101 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.488110 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: E1004 04:34:06.499241 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:06Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:06 crc kubenswrapper[4992]: E1004 04:34:06.499354 4992 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.500968 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.501002 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.501011 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.501025 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.501034 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.603444 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.603480 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.603493 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.603510 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.603523 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.705344 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.705407 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.705419 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.705436 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.705450 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.808566 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.808629 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.808645 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.808676 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.808700 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.910822 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.910883 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.910898 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.910922 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:06 crc kubenswrapper[4992]: I1004 04:34:06.910941 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:06Z","lastTransitionTime":"2025-10-04T04:34:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.013597 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.013641 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.013651 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.013668 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.013681 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.116303 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.116377 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.116398 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.116422 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.116438 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.219629 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.219682 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.219695 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.219714 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.219728 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.319085 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.319087 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.319109 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:07 crc kubenswrapper[4992]: E1004 04:34:07.319516 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:07 crc kubenswrapper[4992]: E1004 04:34:07.319266 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:07 crc kubenswrapper[4992]: E1004 04:34:07.319633 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.321703 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.321738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.321753 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.321769 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.321779 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.424554 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.424586 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.424599 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.424615 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.424627 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.526922 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.526981 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.526995 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.527015 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.527027 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.629729 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.629784 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.629798 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.629820 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.629833 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.732814 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.732859 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.732872 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.732890 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.732903 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.835870 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.835963 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.835978 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.836005 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.836022 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.937864 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.937920 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.937931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.937946 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:07 crc kubenswrapper[4992]: I1004 04:34:07.937958 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:07Z","lastTransitionTime":"2025-10-04T04:34:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.040920 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.040967 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.040977 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.040995 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.041007 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.144078 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.144135 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.144153 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.144178 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.144196 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.246688 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.246730 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.246741 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.246756 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.246769 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.318553 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:08 crc kubenswrapper[4992]: E1004 04:34:08.318894 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.348731 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.348770 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.348781 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.348799 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.348811 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.451822 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.451885 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.451905 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.451930 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.451948 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.554444 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.554476 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.554485 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.554501 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.554511 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.657169 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.657211 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.657225 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.657240 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.657251 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.759907 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.759943 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.759955 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.759972 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.759984 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.863801 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.863854 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.863900 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.863931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.863997 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.966463 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.966526 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.966543 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.966572 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:08 crc kubenswrapper[4992]: I1004 04:34:08.966591 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:08Z","lastTransitionTime":"2025-10-04T04:34:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.069529 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.069590 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.069604 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.069623 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.069635 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.173170 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.173207 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.173218 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.173234 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.173244 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.276823 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.276884 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.276897 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.276920 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.276936 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.319141 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.319191 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.319287 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:09 crc kubenswrapper[4992]: E1004 04:34:09.319464 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:09 crc kubenswrapper[4992]: E1004 04:34:09.319575 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:09 crc kubenswrapper[4992]: E1004 04:34:09.319719 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.379955 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.379999 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.380010 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.380026 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.380038 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.482922 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.482959 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.482970 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.482988 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.483002 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.586115 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.586179 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.586194 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.586219 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.586236 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.688670 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.688738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.688931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.688958 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.688973 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.791423 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.791483 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.791496 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.791515 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.791530 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.894787 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.894841 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.894855 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.894873 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.894885 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.997432 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.997495 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.997507 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.997531 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:09 crc kubenswrapper[4992]: I1004 04:34:09.997546 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:09Z","lastTransitionTime":"2025-10-04T04:34:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.100404 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.100475 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.100489 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.100511 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.100527 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.203489 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.203560 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.203576 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.203599 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.203614 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.305973 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.306029 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.306039 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.306060 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.306071 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.318790 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:10 crc kubenswrapper[4992]: E1004 04:34:10.318983 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.409499 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.409572 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.409591 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.409621 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.409642 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.512675 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.512737 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.512755 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.512780 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.512796 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.615851 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.615900 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.615910 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.615931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.615942 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.718833 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.718897 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.718912 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.718938 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.718956 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.821514 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.821588 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.821605 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.821636 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.821653 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.924946 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.925026 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.925051 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.925085 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:10 crc kubenswrapper[4992]: I1004 04:34:10.925107 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:10Z","lastTransitionTime":"2025-10-04T04:34:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.027680 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.027765 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.027791 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.027817 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.027833 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.130499 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.130528 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.130536 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.130550 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.130560 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.232796 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.232831 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.232841 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.232858 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.232869 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.318841 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:11 crc kubenswrapper[4992]: E1004 04:34:11.319066 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.319485 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:11 crc kubenswrapper[4992]: E1004 04:34:11.319670 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.319746 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:11 crc kubenswrapper[4992]: E1004 04:34:11.319928 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.335439 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.335493 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.335515 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.335544 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.335567 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.438286 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.438409 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.438429 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.438451 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.438466 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.540878 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.540937 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.540956 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.540983 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.541003 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.644020 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.644084 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.644109 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.644136 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.644159 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.747835 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.747884 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.747894 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.747914 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.747925 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.851354 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.851437 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.851467 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.851492 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.851510 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.954824 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.954871 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.954882 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.954906 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:11 crc kubenswrapper[4992]: I1004 04:34:11.954925 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:11Z","lastTransitionTime":"2025-10-04T04:34:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.058314 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.058424 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.058454 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.058478 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.058491 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.161179 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.161230 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.161247 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.161264 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.161275 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.264883 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.264939 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.264952 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.264973 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.264986 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.319270 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:12 crc kubenswrapper[4992]: E1004 04:34:12.319539 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.368082 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.368114 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.368123 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.368139 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.368151 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.471511 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.471670 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.471689 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.471710 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.471724 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.575169 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.575213 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.575226 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.575246 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.575279 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.678620 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.678672 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.678685 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.678701 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.678712 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.781151 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.781233 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.781264 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.781299 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.781324 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.884012 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.884073 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.884093 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.884120 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.884138 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.988260 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.988312 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.988329 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.988353 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:12 crc kubenswrapper[4992]: I1004 04:34:12.988405 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:12Z","lastTransitionTime":"2025-10-04T04:34:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.091186 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.091218 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.091226 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.091241 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.091251 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.197301 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.197347 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.197385 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.197405 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.197422 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.300662 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.300742 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.300752 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.300806 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.300816 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.319609 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.319624 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:13 crc kubenswrapper[4992]: E1004 04:34:13.319830 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.319608 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:13 crc kubenswrapper[4992]: E1004 04:34:13.319916 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:13 crc kubenswrapper[4992]: E1004 04:34:13.320070 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.403583 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.403662 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.403683 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.403738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.403756 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.507516 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.507580 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.507590 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.507605 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.507616 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.610003 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.610049 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.610064 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.610085 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.610100 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.712800 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.712843 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.712855 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.712882 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.712895 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.815659 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.815712 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.815726 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.815744 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.815755 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.919797 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.919850 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.919861 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.919880 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:13 crc kubenswrapper[4992]: I1004 04:34:13.919893 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:13Z","lastTransitionTime":"2025-10-04T04:34:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.023042 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.023094 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.023113 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.023136 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.023155 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.126450 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.126512 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.126524 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.126547 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.126562 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.228835 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.228871 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.228879 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.228892 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.228901 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.319097 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:14 crc kubenswrapper[4992]: E1004 04:34:14.319247 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.331475 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.331513 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.331525 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.331542 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.331553 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.331956 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.334692 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.345284 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.357766 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.370420 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.381697 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.392684 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.412781 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.426536 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.433433 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.433473 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.433483 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.433500 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.433511 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.439179 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.451337 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.459939 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.471691 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:03Z\\\",\\\"message\\\":\\\"2025-10-04T04:33:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e\\\\n2025-10-04T04:33:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e to /host/opt/cni/bin/\\\\n2025-10-04T04:33:18Z [verbose] multus-daemon started\\\\n2025-10-04T04:33:18Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:34:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.484114 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.499906 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.510904 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.525084 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.537267 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.537321 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.537336 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.537439 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.537464 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.546864 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.565483 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:14Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.642281 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.642323 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.642336 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.642352 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.642377 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.744418 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.744456 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.744466 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.744482 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.744493 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.846599 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.846634 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.846642 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.846657 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.846666 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.949557 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.949630 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.949646 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.949679 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:14 crc kubenswrapper[4992]: I1004 04:34:14.949696 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:14Z","lastTransitionTime":"2025-10-04T04:34:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.052058 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.052118 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.052128 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.052145 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.052176 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.154713 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.154742 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.154750 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.154764 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.154774 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.256539 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.256575 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.256584 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.256597 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.256607 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.318482 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.318513 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.318540 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:15 crc kubenswrapper[4992]: E1004 04:34:15.318828 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:15 crc kubenswrapper[4992]: E1004 04:34:15.318950 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:15 crc kubenswrapper[4992]: E1004 04:34:15.319007 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.319097 4992 scope.go:117] "RemoveContainer" containerID="503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.359185 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.359249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.359266 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.359298 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.359320 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.462128 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.462187 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.462200 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.462216 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.462228 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.564612 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.564636 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.564644 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.564657 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.564666 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.666598 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.666647 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.666659 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.666676 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.666688 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.770796 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.770831 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.770841 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.770856 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.770867 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.790523 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/2.log" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.793395 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.793735 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.805728 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50dcc48c-93f4-4187-86e3-305476612334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://919162be30623317a3e81e1435503638a2c7ae724ca46c797213c34cefe590c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efa25998cda3065a1e3634dfc72257edd52cc51ea13c97460b2d5e322a0a6e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efa25998cda3065a1e3634dfc72257edd52cc51ea13c97460b2d5e322a0a6e40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.821685 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.839278 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.854647 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.870979 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.872945 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.872978 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.872990 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.873007 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.873018 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.892315 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.915417 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.934424 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.947156 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.960334 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.974563 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:03Z\\\",\\\"message\\\":\\\"2025-10-04T04:33:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e\\\\n2025-10-04T04:33:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e to /host/opt/cni/bin/\\\\n2025-10-04T04:33:18Z [verbose] multus-daemon started\\\\n2025-10-04T04:33:18Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:34:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.975514 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.975541 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.975550 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.975565 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.975574 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:15Z","lastTransitionTime":"2025-10-04T04:34:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:15 crc kubenswrapper[4992]: I1004 04:34:15.989025 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.001086 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:15Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.013106 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.025866 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.040014 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.058707 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.071955 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.079040 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.079084 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.079097 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.079113 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.079125 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.085085 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.181232 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.181264 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.181272 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.181288 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.181297 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.283837 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.283873 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.283883 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.283898 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.283908 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.318783 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:16 crc kubenswrapper[4992]: E1004 04:34:16.318999 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.386138 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.386175 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.386187 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.386203 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.386219 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.488810 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.488871 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.488890 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.488917 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.488936 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.592252 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.592298 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.592308 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.592553 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.592563 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.675864 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.675956 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.675983 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.676013 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.676069 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: E1004 04:34:16.697591 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.703029 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.703067 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.703077 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.703095 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.703104 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: E1004 04:34:16.717219 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.722687 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.722800 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.722824 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.722850 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.722870 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: E1004 04:34:16.741258 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.744754 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.744791 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.744803 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.744820 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.744833 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: E1004 04:34:16.756505 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.759558 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.759594 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.759603 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.759616 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.759625 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: E1004 04:34:16.770470 4992 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"0b6f8901-3900-4293-9527-60030bed2cef\\\",\\\"systemUUID\\\":\\\"e9142861-7515-4077-bd2a-8f44fe92e099\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: E1004 04:34:16.770581 4992 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.771804 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.771838 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.771849 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.771872 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.771885 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.798240 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/3.log" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.798927 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/2.log" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.801726 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" exitCode=1 Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.801775 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.801829 4992 scope.go:117] "RemoveContainer" containerID="503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.802434 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:34:16 crc kubenswrapper[4992]: E1004 04:34:16.802569 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.823113 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.848214 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.868943 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.878947 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.878980 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.878989 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.879004 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.879013 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.888726 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.901060 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.911173 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.920807 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50dcc48c-93f4-4187-86e3-305476612334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://919162be30623317a3e81e1435503638a2c7ae724ca46c797213c34cefe590c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efa25998cda3065a1e3634dfc72257edd52cc51ea13c97460b2d5e322a0a6e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efa25998cda3065a1e3634dfc72257edd52cc51ea13c97460b2d5e322a0a6e40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.936555 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.951062 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.963809 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.973517 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.981507 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.981719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.981738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.981754 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.981765 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:16Z","lastTransitionTime":"2025-10-04T04:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:16 crc kubenswrapper[4992]: I1004 04:34:16.985567 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:03Z\\\",\\\"message\\\":\\\"2025-10-04T04:33:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e\\\\n2025-10-04T04:33:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e to /host/opt/cni/bin/\\\\n2025-10-04T04:33:18Z [verbose] multus-daemon started\\\\n2025-10-04T04:33:18Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:34:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:16Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.002191 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.013167 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.021756 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.032699 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.045267 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.071957 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://503621a436ba7d1aa64431bd25c54ad9f094752bd34a533b56110b63b42e6020\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:33:47Z\\\",\\\"message\\\":\\\"ations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 04:33:47.200128 6676 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nF1004 04:33:47.200131 6676 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:33:47Z is after 2025-08-24T17:21:41Z]\\\\nI1004 04:33:47.200141 6676 services_controller.go:356] Processing sync for service ope\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1004 04:34:16.273496 7067 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 04:34:16.273614 7067 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 04:34:16.273632 7067 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 04:34:16.273644 7067 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 04:34:16.273683 7067 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:34:16.273960 7067 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1004 04:34:16.274080 7067 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 04:34:16.274437 7067 factory.go:656] Stopping watch factory\\\\nI1004 04:34:16.274466 7067 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:34:16.274488 7067 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 04:34:16.274554 7067 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.083677 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.083718 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.083727 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.083742 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.083751 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.089027 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.186524 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.186564 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.186573 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.186589 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.186599 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.288904 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.288985 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.288995 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.289014 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.289034 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.319479 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.319525 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.319484 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:17 crc kubenswrapper[4992]: E1004 04:34:17.319665 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:17 crc kubenswrapper[4992]: E1004 04:34:17.319782 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:17 crc kubenswrapper[4992]: E1004 04:34:17.319884 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.391087 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.391128 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.391138 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.391152 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.391163 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.494525 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.494633 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.494653 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.494728 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.496115 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.599514 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.599556 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.599569 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.599587 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.599599 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.702142 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.702186 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.702197 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.702215 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.702228 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.804976 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.805018 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.805026 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.805040 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.805052 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.806213 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/3.log" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.809760 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:34:17 crc kubenswrapper[4992]: E1004 04:34:17.809898 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.831984 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3535aa0-52ec-4431-b34e-868cf96e00c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b8f207e2696c9ba6920d26a46012fc62e92ba4539b9f65f3fdb1a18e463506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://300e107e89310012fd290650e657612a97374af857f193acc3144fccef146053\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7a2fc5fd5ca29078bbed2394dcdded4ebead347d125cfb53bd3a020ee825351\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69e83a3a06d5abbd4c81bfdcf45213886f47bc6798988a4b5f1c67e38621aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a417f38741bde30ab52cbc20b3e7ceadc578fe4d54a74af084226a3a1478a4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aead8c6ea84e109c16133abc368d57e564e9af74a0e26e1ff01a055305862b6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bdd19d1f3245e21f0a96335c948f1becaf69f6cab2372e9025ed773927f44ea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6c5c605f7a41d5f7cf7d3d999e0409f8763ac29dedaa4d816e4e24b029b637ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.845472 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09eccc5558b1d69124a782fd4b60f14005d743904e5b95f8119ec9481323d1fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.859809 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.871400 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.880930 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-bckw6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"361f0cb6-df75-48b8-916b-4f6e19bd8834\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9348e03c315f75c038e74a4e5eb99e213e9b21b5b469fb03d40f46f87a1cac0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hdh7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-bckw6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.898917 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2b6879c5-46f4-4d87-959b-03dff08cff4c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:34:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:03Z\\\",\\\"message\\\":\\\"2025-10-04T04:33:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e\\\\n2025-10-04T04:33:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_905cfaf5-d4a4-4ca1-8a55-8048a57b835e to /host/opt/cni/bin/\\\\n2025-10-04T04:33:18Z [verbose] multus-daemon started\\\\n2025-10-04T04:33:18Z [verbose] Readiness Indicator file check\\\\n2025-10-04T04:34:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:16Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:34:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9j29w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.907399 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.907428 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.907436 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.907450 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.907460 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:17Z","lastTransitionTime":"2025-10-04T04:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.909960 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52d16a70-184d-43ad-8f78-b0f2de70b259\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c473bfec3d2ec1b03618c38fc034574555496288a055219af6dae50da02b5c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06cf3d5618ae7684afd4bc6b57032a41b0dc02a321dd5018cd7eb32ee6727857\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad8cafbcbe3ef7043a25a63842ef6415b2f616cbc5df4609077d2f88dd82185d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f53f23ec7f98377b7f2b8d9dc1e907647ae3f0eaf392fed0d56969353f7f190f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.921059 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"723f0d8f-8278-4fbf-b41b-8735c9fe4e6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6b2dc5ccda3953c5b5164569acfda871449da1edf264a968c54dd312944ccfce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acdde6664dd37dd80b05dc073a2b3c2a2b96d7f9666d39ac4bf7fa5a37451752\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e80b1e38d42bc62b19246459e7342d73d62392c5dd6dbd7e3b365e1c97d0f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fc18ad22e1e9420169c59f27f332eea787ba8c77fa424642d38cc0fd977781dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.930715 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ncbdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad0f7c37-e53f-4a34-8da9-575e78b32a95\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c4d3d3c5ba8de542af957c11ee0e863d24f16faa3464a7fb63c3ccb63c570b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9f9fp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ncbdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.942237 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://156c222edf35871aad6c466d91664da4493d9815e485857c5c5060f7a482a4f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11445f81abe658923dd66332fd56c2426e518cf3812e55984b7688b21d83003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.955240 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be8a9bf6-a143-43ab-8f47-f3a72a2e902f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d259df43642591ec4841387a1f951e1cf737a2cf8fc2b16fe5ba8c2dcb666263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f90d5e0895f0096b19a01893753956f9f828645e1a65a1b69584c563e6cc7b9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://738a34df4b7929b653d73e14708d03f6ba2403503163a2df9ff365868a1ef084\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20d42f969bb535236969aef072edcbe4a4056d9631e855ecabcdb82e512402a4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af6545cb58cff1ba14ce92cdc697f070c11e0179eb90eeda838ef5f2e6a8c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6017ab7a0609a58e817739fed7155268829ba1f3323fe20a9e343800973bd6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5962a145868d72b33b95bbfcdc9a6ef9c17dc519de63879fb88f5ed08cd33460\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7b2p7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wp9cz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.973544 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T04:34:16Z\\\",\\\"message\\\":\\\"ory.go:160\\\\nI1004 04:34:16.273496 7067 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 04:34:16.273614 7067 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1004 04:34:16.273632 7067 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1004 04:34:16.273644 7067 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 04:34:16.273683 7067 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 04:34:16.273960 7067 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1004 04:34:16.274080 7067 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 04:34:16.274437 7067 factory.go:656] Stopping watch factory\\\\nI1004 04:34:16.274466 7067 ovnkube.go:599] Stopped ovnkube\\\\nI1004 04:34:16.274488 7067 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1004 04:34:16.274554 7067 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:34:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b7qm9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fnr76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:17 crc kubenswrapper[4992]: I1004 04:34:17.989339 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7018a9aa-127b-4f97-a117-6c69bdeb4803\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00cd44f3aa3ac74330fff2d8b50aed3dc7208a8a51f6f85755028f14e6839524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://925333e1663e7535414dcfae0fdb54025dd7d0905f652a8487038a882ccecc55\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01fb3de2a6734bbd71315303c86861c8f2bef7287ea304cfd32f01bb58a4553f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c23c5ef899010fe80fc1eeb643adee2aa4a0e799cd3ed78445f02c9fd912c8d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1764cad303a04e2ef59999fac4092bdfe716ddb0b3c163a2452fa1e82c2019a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"32:58 +0000 UTC to 2025-11-03 04:32:59 +0000 UTC (now=2025-10-04 04:33:14.846773239 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846832 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1004 04:33:14.846861 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-861087537/tls.crt::/tmp/serving-cert-861087537/tls.key\\\\\\\"\\\\nI1004 04:33:14.846877 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1004 04:33:14.846965 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1004 04:33:14.846970 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759552380\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759552379\\\\\\\\\\\\\\\" (2025-10-04 03:32:59 +0000 UTC to 2026-10-04 03:32:59 +0000 UTC (now=2025-10-04 04:33:14.846951704 +0000 UTC))\\\\\\\"\\\\nI1004 04:33:14.846993 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 04:33:14.847016 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 04:33:14.847059 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI1004 04:33:14.846982 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1004 04:33:14.851707 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.851711 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nI1004 04:33:14.852090 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243\\\\nF1004 04:33:14.856887 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91bffc84216de2b6da8fb98490134212ac2732304b2fce19a42b38bb9898c109\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c42b28d8044864714c0a14b06c18d5da630f339fa327643d447f50336ae9ce7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:17Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.010397 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.010445 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.010457 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.010474 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.010484 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.011838 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb369069dc836e59c7193bea9d1f8abad21cea0ea25d6829291d24629b58d723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.022449 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50dcc48c-93f4-4187-86e3-305476612334\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://919162be30623317a3e81e1435503638a2c7ae724ca46c797213c34cefe590c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:32:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://efa25998cda3065a1e3634dfc72257edd52cc51ea13c97460b2d5e322a0a6e40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://efa25998cda3065a1e3634dfc72257edd52cc51ea13c97460b2d5e322a0a6e40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T04:32:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T04:32:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:32:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.035178 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.048761 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6d4bf54-7f00-4c33-830c-8368dd9678df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e7ea157ad771cf81991279afd0d027acb161578ed38fc16ab0f5b33922d9f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q5c79\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p5l9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.061104 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47bf1eb8-79e2-46d7-b4b0-1b0947872fa0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://292487c9cd9804cf717d0f3f8ee7a6dc67d4a0201e80e5031af215c1478f7856\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07cbb7bd1cc1911455beb0b077a4076affc2b48f977884ba3b3edaefdb7f1fe7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T04:33:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fc772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:28Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ft6s9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.071909 4992 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-6jlms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ef62768-7ed6-4ce3-9d34-6e6743681405\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T04:33:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nhsvj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T04:33:29Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-6jlms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T04:34:18Z is after 2025-08-24T17:21:41Z" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.113225 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.113271 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.113318 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.113335 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.113349 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.216197 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.216255 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.216268 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.216285 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.216296 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.318499 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:18 crc kubenswrapper[4992]: E1004 04:34:18.318608 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.318788 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.318830 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.318842 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.318858 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.318869 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.421227 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.421287 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.421305 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.421328 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.421348 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.524253 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.524297 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.524306 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.524320 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.524329 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.627758 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.627810 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.627822 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.627841 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.627856 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.730256 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.730301 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.730313 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.730332 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.730553 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.833948 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.833990 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.834000 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.834016 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.834027 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.935886 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.935934 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.935944 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.935961 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:18 crc kubenswrapper[4992]: I1004 04:34:18.935972 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:18Z","lastTransitionTime":"2025-10-04T04:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.037778 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.037819 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.037832 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.037848 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.037860 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.139931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.139964 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.139972 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.139990 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.140000 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.149690 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.149857 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.149835654 +0000 UTC m=+156.997511172 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.242280 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.242336 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.242347 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.242385 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.242394 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.251105 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.251157 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.251194 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251204 4992 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251270 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.251251303 +0000 UTC m=+157.098926841 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251296 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251310 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251321 4992 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.251210 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251384 4992 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251399 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.251350196 +0000 UTC m=+157.099025724 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251490 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.251466309 +0000 UTC m=+157.099141827 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251503 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251542 4992 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251555 4992 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.251623 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.251604753 +0000 UTC m=+157.099280291 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.318639 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.318690 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.318650 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.318760 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.318840 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:19 crc kubenswrapper[4992]: E1004 04:34:19.318918 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.345557 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.345598 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.345609 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.345627 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.345639 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.448042 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.448102 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.448115 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.448131 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.448141 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.550692 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.550738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.550752 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.550772 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.550787 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.654651 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.654691 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.654702 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.654719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.654731 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.757734 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.757771 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.757780 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.757794 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.757803 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.860691 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.860799 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.860815 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.860839 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.860855 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.964709 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.964758 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.964769 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.964786 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:19 crc kubenswrapper[4992]: I1004 04:34:19.964798 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:19Z","lastTransitionTime":"2025-10-04T04:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.067834 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.067876 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.067888 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.067904 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.067916 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.169721 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.169791 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.169807 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.169823 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.169834 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.272488 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.272528 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.272540 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.272558 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.272570 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.318957 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:20 crc kubenswrapper[4992]: E1004 04:34:20.319169 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.375823 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.375889 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.375909 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.375934 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.375952 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.478738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.478851 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.478886 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.478923 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.478942 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.582168 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.582238 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.582249 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.582265 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.582280 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.684634 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.684682 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.684701 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.684722 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.684737 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.788321 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.788405 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.788420 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.788441 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.788456 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.890921 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.891321 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.891334 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.891351 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.891386 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.994466 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.994532 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.994555 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.994589 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:20 crc kubenswrapper[4992]: I1004 04:34:20.994612 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:20Z","lastTransitionTime":"2025-10-04T04:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.098041 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.098184 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.098207 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.098237 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.098255 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.201659 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.201695 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.201704 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.201719 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.201728 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.304464 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.304503 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.304514 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.304531 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.304546 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.319309 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.319382 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.319329 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:21 crc kubenswrapper[4992]: E1004 04:34:21.319648 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:21 crc kubenswrapper[4992]: E1004 04:34:21.319762 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:21 crc kubenswrapper[4992]: E1004 04:34:21.319874 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.406936 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.406983 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.406995 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.407013 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.407025 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.509332 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.509402 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.509413 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.509428 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.509439 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.612808 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.612845 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.612859 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.612876 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.612888 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.715529 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.715581 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.715596 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.715618 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.715630 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.818664 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.818716 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.818733 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.818752 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.818768 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.920516 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.920565 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.920576 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.920596 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:21 crc kubenswrapper[4992]: I1004 04:34:21.920609 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:21Z","lastTransitionTime":"2025-10-04T04:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.023429 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.023473 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.023488 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.023504 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.023514 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.126569 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.126618 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.126634 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.126655 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.126671 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.229057 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.229110 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.229125 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.229146 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.229162 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.319014 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:22 crc kubenswrapper[4992]: E1004 04:34:22.319138 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.330795 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.330824 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.330832 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.330842 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.330851 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.433689 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.433726 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.433737 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.433761 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.433783 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.536124 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.536169 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.536181 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.536202 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.536219 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.638291 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.638412 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.638438 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.638469 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.638492 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.741351 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.741430 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.741443 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.741470 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.741484 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.843873 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.843926 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.843937 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.843963 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.843979 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.947097 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.947143 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.947152 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.947171 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:22 crc kubenswrapper[4992]: I1004 04:34:22.947183 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:22Z","lastTransitionTime":"2025-10-04T04:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.050661 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.050737 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.050755 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.050778 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.050791 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.153681 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.153729 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.153743 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.153760 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.153774 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.257431 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.257484 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.257497 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.257522 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.257536 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.318456 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.318555 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:23 crc kubenswrapper[4992]: E1004 04:34:23.318667 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.318704 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:23 crc kubenswrapper[4992]: E1004 04:34:23.318870 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:23 crc kubenswrapper[4992]: E1004 04:34:23.318920 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.359694 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.359738 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.359749 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.359766 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.359777 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.463017 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.463080 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.463094 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.463114 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.463126 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.566562 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.566647 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.566670 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.566701 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.566722 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.668951 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.668994 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.669005 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.669021 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.669032 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.771295 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.771340 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.771352 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.771405 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.771418 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.873080 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.873114 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.873122 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.873139 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.873155 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.974949 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.974986 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.974997 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.975012 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:23 crc kubenswrapper[4992]: I1004 04:34:23.975021 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:23Z","lastTransitionTime":"2025-10-04T04:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.078195 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.078265 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.078288 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.078315 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.078332 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.183148 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.183189 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.183198 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.183213 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.183225 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.286274 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.286324 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.286342 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.286376 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.286395 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.318688 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:24 crc kubenswrapper[4992]: E1004 04:34:24.318827 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.363853 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.36383847 podStartE2EDuration="1m10.36383847s" podCreationTimestamp="2025-10-04 04:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.351598149 +0000 UTC m=+98.199273627" watchObservedRunningTime="2025-10-04 04:34:24.36383847 +0000 UTC m=+98.211513938" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.375931 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ft6s9" podStartSLOduration=68.375913717 podStartE2EDuration="1m8.375913717s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.375753652 +0000 UTC m=+98.223429130" watchObservedRunningTime="2025-10-04 04:34:24.375913717 +0000 UTC m=+98.223589185" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.389286 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.389322 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.389333 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.389346 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.389372 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.398064 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.398043407 podStartE2EDuration="10.398043407s" podCreationTimestamp="2025-10-04 04:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.398003706 +0000 UTC m=+98.245679184" watchObservedRunningTime="2025-10-04 04:34:24.398043407 +0000 UTC m=+98.245718875" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.422172 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podStartSLOduration=69.42215029 podStartE2EDuration="1m9.42215029s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.420528287 +0000 UTC m=+98.268203755" watchObservedRunningTime="2025-10-04 04:34:24.42215029 +0000 UTC m=+98.269825758" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.455703 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-bckw6" podStartSLOduration=69.455682249 podStartE2EDuration="1m9.455682249s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.442509434 +0000 UTC m=+98.290184902" watchObservedRunningTime="2025-10-04 04:34:24.455682249 +0000 UTC m=+98.303357717" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.455901 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vhq64" podStartSLOduration=69.455897075 podStartE2EDuration="1m9.455897075s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.455505685 +0000 UTC m=+98.303181153" watchObservedRunningTime="2025-10-04 04:34:24.455897075 +0000 UTC m=+98.303572533" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.489811 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=68.489792934 podStartE2EDuration="1m8.489792934s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.488921581 +0000 UTC m=+98.336597049" watchObservedRunningTime="2025-10-04 04:34:24.489792934 +0000 UTC m=+98.337468402" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.492137 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.492186 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.492199 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.492220 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.492234 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.553662 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-wp9cz" podStartSLOduration=69.553644819 podStartE2EDuration="1m9.553644819s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.553388322 +0000 UTC m=+98.401063810" watchObservedRunningTime="2025-10-04 04:34:24.553644819 +0000 UTC m=+98.401320287" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.595137 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.595175 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.595185 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.595197 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.595206 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.597212 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=42.597194731 podStartE2EDuration="42.597194731s" podCreationTimestamp="2025-10-04 04:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.59675734 +0000 UTC m=+98.444432808" watchObservedRunningTime="2025-10-04 04:34:24.597194731 +0000 UTC m=+98.444870190" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.597845 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.597838438 podStartE2EDuration="1m4.597838438s" podCreationTimestamp="2025-10-04 04:33:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.585885905 +0000 UTC m=+98.433561373" watchObservedRunningTime="2025-10-04 04:34:24.597838438 +0000 UTC m=+98.445513906" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.697969 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.698012 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.698023 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.698040 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.698052 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.800315 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.800404 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.800419 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.800435 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.800444 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.903180 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.903217 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.903227 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.903246 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:24 crc kubenswrapper[4992]: I1004 04:34:24.903259 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:24Z","lastTransitionTime":"2025-10-04T04:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.005240 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.005580 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.005705 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.005805 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.005887 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.108190 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.108246 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.108260 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.108277 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.108289 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.210858 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.210903 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.210916 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.210931 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.210957 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.313386 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.313417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.313426 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.313438 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.313447 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.318847 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.318856 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.318856 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:25 crc kubenswrapper[4992]: E1004 04:34:25.319335 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:25 crc kubenswrapper[4992]: E1004 04:34:25.319411 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:25 crc kubenswrapper[4992]: E1004 04:34:25.319525 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.423065 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.423341 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.423438 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.423516 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.423589 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.525625 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.525665 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.525677 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.525694 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.525705 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.627936 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.627970 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.627981 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.627996 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.628006 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.730406 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.730448 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.730458 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.730474 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.730484 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.832428 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.832467 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.832477 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.832492 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.832502 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.934997 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.935038 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.935051 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.935064 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:25 crc kubenswrapper[4992]: I1004 04:34:25.935072 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:25Z","lastTransitionTime":"2025-10-04T04:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.036907 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.036973 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.036992 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.037017 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.037035 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.139976 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.140013 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.140021 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.140037 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.140048 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.242700 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.242743 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.242753 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.242767 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.242779 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.318719 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:26 crc kubenswrapper[4992]: E1004 04:34:26.318872 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.344561 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.344602 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.344613 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.344630 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.344641 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.446452 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.446486 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.446494 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.446506 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.446514 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.548849 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.548927 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.548941 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.548957 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.548967 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.651550 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.651930 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.652005 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.652074 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.652156 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.755066 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.755116 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.755131 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.755153 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.755166 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.857455 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.857541 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.857566 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.857599 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.857621 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.959315 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.959561 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.959636 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.959718 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:26 crc kubenswrapper[4992]: I1004 04:34:26.959777 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:26Z","lastTransitionTime":"2025-10-04T04:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.062314 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.062400 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.062417 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.062440 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.062457 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:27Z","lastTransitionTime":"2025-10-04T04:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.087818 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.087903 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.087927 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.087960 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.087987 4992 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T04:34:27Z","lastTransitionTime":"2025-10-04T04:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.143935 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ncbdl" podStartSLOduration=73.143910508 podStartE2EDuration="1m13.143910508s" podCreationTimestamp="2025-10-04 04:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:24.608299463 +0000 UTC m=+98.455974931" watchObservedRunningTime="2025-10-04 04:34:27.143910508 +0000 UTC m=+100.991586026" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.144974 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj"] Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.145619 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.148662 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.148772 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.149033 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.151835 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.248586 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9a8d022-78b4-41ae-b023-e82e79cca506-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.248765 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9a8d022-78b4-41ae-b023-e82e79cca506-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.248838 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c9a8d022-78b4-41ae-b023-e82e79cca506-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.248866 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c9a8d022-78b4-41ae-b023-e82e79cca506-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.248941 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c9a8d022-78b4-41ae-b023-e82e79cca506-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.319006 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.319084 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:27 crc kubenswrapper[4992]: E1004 04:34:27.319164 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.319096 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:27 crc kubenswrapper[4992]: E1004 04:34:27.319419 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:27 crc kubenswrapper[4992]: E1004 04:34:27.319452 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.349678 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9a8d022-78b4-41ae-b023-e82e79cca506-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.349752 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9a8d022-78b4-41ae-b023-e82e79cca506-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.349783 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c9a8d022-78b4-41ae-b023-e82e79cca506-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.349812 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c9a8d022-78b4-41ae-b023-e82e79cca506-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.349868 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c9a8d022-78b4-41ae-b023-e82e79cca506-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.349967 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c9a8d022-78b4-41ae-b023-e82e79cca506-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.350023 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c9a8d022-78b4-41ae-b023-e82e79cca506-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.351229 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c9a8d022-78b4-41ae-b023-e82e79cca506-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.357249 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9a8d022-78b4-41ae-b023-e82e79cca506-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.374665 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9a8d022-78b4-41ae-b023-e82e79cca506-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dkjj\" (UID: \"c9a8d022-78b4-41ae-b023-e82e79cca506\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.466900 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.839523 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" event={"ID":"c9a8d022-78b4-41ae-b023-e82e79cca506","Type":"ContainerStarted","Data":"73813faf9cbf1a307209d6d6030d7363afb9cf10b14cbc671bd55da7bb2a816f"} Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.839610 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" event={"ID":"c9a8d022-78b4-41ae-b023-e82e79cca506","Type":"ContainerStarted","Data":"08c09bdfea0dd803fa5de0ce4eb255de3e78330e5ac3238d0ecb1f66e6849e24"} Oct 04 04:34:27 crc kubenswrapper[4992]: I1004 04:34:27.855800 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dkjj" podStartSLOduration=72.855765552 podStartE2EDuration="1m12.855765552s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:34:27.853993346 +0000 UTC m=+101.701668814" watchObservedRunningTime="2025-10-04 04:34:27.855765552 +0000 UTC m=+101.703441020" Oct 04 04:34:28 crc kubenswrapper[4992]: I1004 04:34:28.319068 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:28 crc kubenswrapper[4992]: E1004 04:34:28.319198 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:29 crc kubenswrapper[4992]: I1004 04:34:29.318398 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:29 crc kubenswrapper[4992]: I1004 04:34:29.318408 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:29 crc kubenswrapper[4992]: I1004 04:34:29.318517 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:29 crc kubenswrapper[4992]: E1004 04:34:29.318731 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:29 crc kubenswrapper[4992]: E1004 04:34:29.318835 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:29 crc kubenswrapper[4992]: E1004 04:34:29.318880 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:30 crc kubenswrapper[4992]: I1004 04:34:30.318535 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:30 crc kubenswrapper[4992]: E1004 04:34:30.318677 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:31 crc kubenswrapper[4992]: I1004 04:34:31.318593 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:31 crc kubenswrapper[4992]: I1004 04:34:31.318658 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:31 crc kubenswrapper[4992]: I1004 04:34:31.318716 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:31 crc kubenswrapper[4992]: E1004 04:34:31.318765 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:31 crc kubenswrapper[4992]: E1004 04:34:31.319012 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:31 crc kubenswrapper[4992]: E1004 04:34:31.319142 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:32 crc kubenswrapper[4992]: I1004 04:34:32.318184 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:32 crc kubenswrapper[4992]: E1004 04:34:32.318281 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:33 crc kubenswrapper[4992]: I1004 04:34:33.318862 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:33 crc kubenswrapper[4992]: E1004 04:34:33.318966 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:33 crc kubenswrapper[4992]: I1004 04:34:33.319014 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:33 crc kubenswrapper[4992]: I1004 04:34:33.319107 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:33 crc kubenswrapper[4992]: E1004 04:34:33.319206 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:33 crc kubenswrapper[4992]: E1004 04:34:33.319947 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:33 crc kubenswrapper[4992]: I1004 04:34:33.320435 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:34:33 crc kubenswrapper[4992]: E1004 04:34:33.320703 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:34:33 crc kubenswrapper[4992]: I1004 04:34:33.717348 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:33 crc kubenswrapper[4992]: E1004 04:34:33.717691 4992 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:33 crc kubenswrapper[4992]: E1004 04:34:33.717786 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs podName:2ef62768-7ed6-4ce3-9d34-6e6743681405 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:37.717757407 +0000 UTC m=+171.565432935 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs") pod "network-metrics-daemon-6jlms" (UID: "2ef62768-7ed6-4ce3-9d34-6e6743681405") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 04:34:34 crc kubenswrapper[4992]: I1004 04:34:34.318808 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:34 crc kubenswrapper[4992]: E1004 04:34:34.320036 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:35 crc kubenswrapper[4992]: I1004 04:34:35.321392 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:35 crc kubenswrapper[4992]: E1004 04:34:35.321615 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:35 crc kubenswrapper[4992]: I1004 04:34:35.322279 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:35 crc kubenswrapper[4992]: E1004 04:34:35.322496 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:35 crc kubenswrapper[4992]: I1004 04:34:35.322346 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:35 crc kubenswrapper[4992]: E1004 04:34:35.322760 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:36 crc kubenswrapper[4992]: I1004 04:34:36.319218 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:36 crc kubenswrapper[4992]: E1004 04:34:36.319476 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:37 crc kubenswrapper[4992]: I1004 04:34:37.318927 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:37 crc kubenswrapper[4992]: E1004 04:34:37.319114 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:37 crc kubenswrapper[4992]: I1004 04:34:37.319304 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:37 crc kubenswrapper[4992]: E1004 04:34:37.319599 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:37 crc kubenswrapper[4992]: I1004 04:34:37.318960 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:37 crc kubenswrapper[4992]: E1004 04:34:37.319788 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:38 crc kubenswrapper[4992]: I1004 04:34:38.318697 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:38 crc kubenswrapper[4992]: E1004 04:34:38.318916 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:39 crc kubenswrapper[4992]: I1004 04:34:39.318463 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:39 crc kubenswrapper[4992]: I1004 04:34:39.318579 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:39 crc kubenswrapper[4992]: E1004 04:34:39.319003 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:39 crc kubenswrapper[4992]: I1004 04:34:39.318658 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:39 crc kubenswrapper[4992]: E1004 04:34:39.319086 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:39 crc kubenswrapper[4992]: E1004 04:34:39.319020 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:40 crc kubenswrapper[4992]: I1004 04:34:40.318186 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:40 crc kubenswrapper[4992]: E1004 04:34:40.318985 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:41 crc kubenswrapper[4992]: I1004 04:34:41.318425 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:41 crc kubenswrapper[4992]: E1004 04:34:41.318683 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:41 crc kubenswrapper[4992]: I1004 04:34:41.318777 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:41 crc kubenswrapper[4992]: E1004 04:34:41.319038 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:41 crc kubenswrapper[4992]: I1004 04:34:41.319461 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:41 crc kubenswrapper[4992]: E1004 04:34:41.319645 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:42 crc kubenswrapper[4992]: I1004 04:34:42.318754 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:42 crc kubenswrapper[4992]: E1004 04:34:42.318894 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:43 crc kubenswrapper[4992]: I1004 04:34:43.318681 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:43 crc kubenswrapper[4992]: E1004 04:34:43.318888 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:43 crc kubenswrapper[4992]: I1004 04:34:43.318691 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:43 crc kubenswrapper[4992]: E1004 04:34:43.319018 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:43 crc kubenswrapper[4992]: I1004 04:34:43.318680 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:43 crc kubenswrapper[4992]: E1004 04:34:43.319105 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:44 crc kubenswrapper[4992]: I1004 04:34:44.318593 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:44 crc kubenswrapper[4992]: E1004 04:34:44.319685 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:45 crc kubenswrapper[4992]: I1004 04:34:45.318464 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:45 crc kubenswrapper[4992]: I1004 04:34:45.318496 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:45 crc kubenswrapper[4992]: I1004 04:34:45.318465 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:45 crc kubenswrapper[4992]: E1004 04:34:45.318787 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:45 crc kubenswrapper[4992]: E1004 04:34:45.318893 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:45 crc kubenswrapper[4992]: E1004 04:34:45.319007 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:46 crc kubenswrapper[4992]: I1004 04:34:46.318862 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:46 crc kubenswrapper[4992]: E1004 04:34:46.318970 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:47 crc kubenswrapper[4992]: I1004 04:34:47.319100 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:47 crc kubenswrapper[4992]: I1004 04:34:47.319171 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:47 crc kubenswrapper[4992]: I1004 04:34:47.319115 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:47 crc kubenswrapper[4992]: E1004 04:34:47.319267 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:47 crc kubenswrapper[4992]: E1004 04:34:47.319413 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:47 crc kubenswrapper[4992]: E1004 04:34:47.319858 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:47 crc kubenswrapper[4992]: I1004 04:34:47.320052 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:34:47 crc kubenswrapper[4992]: E1004 04:34:47.320201 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fnr76_openshift-ovn-kubernetes(2e301ed8-9506-4d6e-bbce-d2ad3817ca36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" Oct 04 04:34:48 crc kubenswrapper[4992]: I1004 04:34:48.318301 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:48 crc kubenswrapper[4992]: E1004 04:34:48.318454 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:49 crc kubenswrapper[4992]: I1004 04:34:49.318159 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:49 crc kubenswrapper[4992]: E1004 04:34:49.318419 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:49 crc kubenswrapper[4992]: I1004 04:34:49.318497 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:49 crc kubenswrapper[4992]: I1004 04:34:49.318584 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:49 crc kubenswrapper[4992]: E1004 04:34:49.318660 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:49 crc kubenswrapper[4992]: E1004 04:34:49.318727 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:50 crc kubenswrapper[4992]: I1004 04:34:50.319212 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:50 crc kubenswrapper[4992]: E1004 04:34:50.319666 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:50 crc kubenswrapper[4992]: I1004 04:34:50.912927 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/1.log" Oct 04 04:34:50 crc kubenswrapper[4992]: I1004 04:34:50.913354 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/0.log" Oct 04 04:34:50 crc kubenswrapper[4992]: I1004 04:34:50.913412 4992 generic.go:334] "Generic (PLEG): container finished" podID="2b6879c5-46f4-4d87-959b-03dff08cff4c" containerID="e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105" exitCode=1 Oct 04 04:34:50 crc kubenswrapper[4992]: I1004 04:34:50.913440 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhq64" event={"ID":"2b6879c5-46f4-4d87-959b-03dff08cff4c","Type":"ContainerDied","Data":"e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105"} Oct 04 04:34:50 crc kubenswrapper[4992]: I1004 04:34:50.913470 4992 scope.go:117] "RemoveContainer" containerID="7bc8ce04307870f47987fff89156da8c93b7efc095489ff8fd3fd9e2ac9d1a79" Oct 04 04:34:50 crc kubenswrapper[4992]: I1004 04:34:50.913861 4992 scope.go:117] "RemoveContainer" containerID="e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105" Oct 04 04:34:50 crc kubenswrapper[4992]: E1004 04:34:50.914038 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vhq64_openshift-multus(2b6879c5-46f4-4d87-959b-03dff08cff4c)\"" pod="openshift-multus/multus-vhq64" podUID="2b6879c5-46f4-4d87-959b-03dff08cff4c" Oct 04 04:34:51 crc kubenswrapper[4992]: I1004 04:34:51.318196 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:51 crc kubenswrapper[4992]: E1004 04:34:51.318699 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:51 crc kubenswrapper[4992]: I1004 04:34:51.318229 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:51 crc kubenswrapper[4992]: E1004 04:34:51.318829 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:51 crc kubenswrapper[4992]: I1004 04:34:51.318196 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:51 crc kubenswrapper[4992]: E1004 04:34:51.318910 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:51 crc kubenswrapper[4992]: I1004 04:34:51.921577 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/1.log" Oct 04 04:34:52 crc kubenswrapper[4992]: I1004 04:34:52.318381 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:52 crc kubenswrapper[4992]: E1004 04:34:52.318478 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:53 crc kubenswrapper[4992]: I1004 04:34:53.318693 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:53 crc kubenswrapper[4992]: I1004 04:34:53.318822 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:53 crc kubenswrapper[4992]: E1004 04:34:53.318861 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:53 crc kubenswrapper[4992]: I1004 04:34:53.318822 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:53 crc kubenswrapper[4992]: E1004 04:34:53.318967 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:53 crc kubenswrapper[4992]: E1004 04:34:53.319077 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:54 crc kubenswrapper[4992]: E1004 04:34:54.261423 4992 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 04:34:54 crc kubenswrapper[4992]: I1004 04:34:54.318572 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:54 crc kubenswrapper[4992]: E1004 04:34:54.319682 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:54 crc kubenswrapper[4992]: E1004 04:34:54.475476 4992 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 04:34:55 crc kubenswrapper[4992]: I1004 04:34:55.318338 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:55 crc kubenswrapper[4992]: I1004 04:34:55.318413 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:55 crc kubenswrapper[4992]: I1004 04:34:55.318498 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:55 crc kubenswrapper[4992]: E1004 04:34:55.318602 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:55 crc kubenswrapper[4992]: E1004 04:34:55.318714 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:55 crc kubenswrapper[4992]: E1004 04:34:55.318805 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:56 crc kubenswrapper[4992]: I1004 04:34:56.318466 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:56 crc kubenswrapper[4992]: E1004 04:34:56.318597 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:57 crc kubenswrapper[4992]: I1004 04:34:57.318735 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:57 crc kubenswrapper[4992]: I1004 04:34:57.318771 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:57 crc kubenswrapper[4992]: I1004 04:34:57.319028 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:57 crc kubenswrapper[4992]: E1004 04:34:57.319085 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:57 crc kubenswrapper[4992]: E1004 04:34:57.318913 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:57 crc kubenswrapper[4992]: E1004 04:34:57.319249 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:58 crc kubenswrapper[4992]: I1004 04:34:58.318978 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:34:58 crc kubenswrapper[4992]: E1004 04:34:58.319145 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:34:59 crc kubenswrapper[4992]: I1004 04:34:59.318376 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:34:59 crc kubenswrapper[4992]: I1004 04:34:59.318431 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:34:59 crc kubenswrapper[4992]: E1004 04:34:59.318525 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:34:59 crc kubenswrapper[4992]: I1004 04:34:59.318402 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:34:59 crc kubenswrapper[4992]: E1004 04:34:59.318603 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:34:59 crc kubenswrapper[4992]: E1004 04:34:59.318669 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:34:59 crc kubenswrapper[4992]: E1004 04:34:59.476615 4992 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 04:35:00 crc kubenswrapper[4992]: I1004 04:35:00.319275 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:00 crc kubenswrapper[4992]: E1004 04:35:00.319547 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:01 crc kubenswrapper[4992]: I1004 04:35:01.318606 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:01 crc kubenswrapper[4992]: I1004 04:35:01.318755 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:01 crc kubenswrapper[4992]: I1004 04:35:01.319112 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:01 crc kubenswrapper[4992]: E1004 04:35:01.319266 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:35:01 crc kubenswrapper[4992]: E1004 04:35:01.319736 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:01 crc kubenswrapper[4992]: E1004 04:35:01.319918 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:01 crc kubenswrapper[4992]: I1004 04:35:01.319964 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:35:01 crc kubenswrapper[4992]: I1004 04:35:01.958098 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/3.log" Oct 04 04:35:01 crc kubenswrapper[4992]: I1004 04:35:01.960953 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerStarted","Data":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} Oct 04 04:35:01 crc kubenswrapper[4992]: I1004 04:35:01.961425 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:35:01 crc kubenswrapper[4992]: I1004 04:35:01.991819 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podStartSLOduration=105.991797881 podStartE2EDuration="1m45.991797881s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:01.991150244 +0000 UTC m=+135.838825742" watchObservedRunningTime="2025-10-04 04:35:01.991797881 +0000 UTC m=+135.839473359" Oct 04 04:35:02 crc kubenswrapper[4992]: I1004 04:35:02.258407 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6jlms"] Oct 04 04:35:02 crc kubenswrapper[4992]: I1004 04:35:02.258558 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:02 crc kubenswrapper[4992]: E1004 04:35:02.258649 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:35:02 crc kubenswrapper[4992]: I1004 04:35:02.318850 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:02 crc kubenswrapper[4992]: E1004 04:35:02.319000 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:03 crc kubenswrapper[4992]: I1004 04:35:03.319083 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:03 crc kubenswrapper[4992]: I1004 04:35:03.319167 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:03 crc kubenswrapper[4992]: E1004 04:35:03.319258 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:03 crc kubenswrapper[4992]: E1004 04:35:03.319444 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:04 crc kubenswrapper[4992]: I1004 04:35:04.319072 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:04 crc kubenswrapper[4992]: I1004 04:35:04.319142 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:04 crc kubenswrapper[4992]: E1004 04:35:04.320380 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:35:04 crc kubenswrapper[4992]: I1004 04:35:04.320438 4992 scope.go:117] "RemoveContainer" containerID="e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105" Oct 04 04:35:04 crc kubenswrapper[4992]: E1004 04:35:04.320473 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:04 crc kubenswrapper[4992]: E1004 04:35:04.477681 4992 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 04:35:04 crc kubenswrapper[4992]: I1004 04:35:04.971419 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/1.log" Oct 04 04:35:04 crc kubenswrapper[4992]: I1004 04:35:04.971502 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhq64" event={"ID":"2b6879c5-46f4-4d87-959b-03dff08cff4c","Type":"ContainerStarted","Data":"f5c27d0d56f2bb47591a59be7ae33f559c7fefcb65ab4b8a0ea45e4c3b25a61c"} Oct 04 04:35:05 crc kubenswrapper[4992]: I1004 04:35:05.319192 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:05 crc kubenswrapper[4992]: I1004 04:35:05.319319 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:05 crc kubenswrapper[4992]: E1004 04:35:05.319425 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:05 crc kubenswrapper[4992]: E1004 04:35:05.319581 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:06 crc kubenswrapper[4992]: I1004 04:35:06.318995 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:06 crc kubenswrapper[4992]: I1004 04:35:06.318995 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:06 crc kubenswrapper[4992]: E1004 04:35:06.319263 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:06 crc kubenswrapper[4992]: E1004 04:35:06.319164 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:35:07 crc kubenswrapper[4992]: I1004 04:35:07.318150 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:07 crc kubenswrapper[4992]: I1004 04:35:07.318221 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:07 crc kubenswrapper[4992]: E1004 04:35:07.318280 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:07 crc kubenswrapper[4992]: E1004 04:35:07.318388 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:08 crc kubenswrapper[4992]: I1004 04:35:08.318671 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:08 crc kubenswrapper[4992]: I1004 04:35:08.318669 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:08 crc kubenswrapper[4992]: E1004 04:35:08.318816 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6jlms" podUID="2ef62768-7ed6-4ce3-9d34-6e6743681405" Oct 04 04:35:08 crc kubenswrapper[4992]: E1004 04:35:08.318944 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 04:35:09 crc kubenswrapper[4992]: I1004 04:35:09.319181 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:09 crc kubenswrapper[4992]: I1004 04:35:09.319235 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:09 crc kubenswrapper[4992]: E1004 04:35:09.319340 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 04:35:09 crc kubenswrapper[4992]: E1004 04:35:09.319466 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 04:35:10 crc kubenswrapper[4992]: I1004 04:35:10.318661 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:10 crc kubenswrapper[4992]: I1004 04:35:10.318732 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:10 crc kubenswrapper[4992]: I1004 04:35:10.321956 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 04:35:10 crc kubenswrapper[4992]: I1004 04:35:10.322197 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 04:35:10 crc kubenswrapper[4992]: I1004 04:35:10.322326 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 04:35:10 crc kubenswrapper[4992]: I1004 04:35:10.322429 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 04:35:11 crc kubenswrapper[4992]: I1004 04:35:11.318395 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:11 crc kubenswrapper[4992]: I1004 04:35:11.318509 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:11 crc kubenswrapper[4992]: I1004 04:35:11.320916 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 04:35:11 crc kubenswrapper[4992]: I1004 04:35:11.320930 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.313160 4992 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.514076 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.514806 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.515168 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fb7m9"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.515713 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.519696 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.520460 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.520872 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.521371 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.525475 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hmx5v"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.526061 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: W1004 04:35:18.526432 4992 reflector.go:561] object-"openshift-route-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.526606 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: E1004 04:35:18.526727 4992 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.527008 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.527118 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.527543 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.528014 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.528458 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pbx8v"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.528567 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.528617 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.528849 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.528967 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.529137 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.529254 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.529663 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.529945 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.530114 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.530455 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.530579 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.530724 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.530800 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.530868 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.530949 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.530957 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.532523 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.532656 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.532782 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.544066 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.544593 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.545482 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cxgpw"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.546021 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.546099 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.546628 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.549988 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.550181 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.550305 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.550478 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.550508 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.550857 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.551032 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.551051 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.551233 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.551171 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.551946 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.552085 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.552263 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.553121 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.554239 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-79tvb"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.554830 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.555282 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.555651 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.556619 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.556845 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.557008 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.557081 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.557266 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.557342 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.557016 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.557625 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.561390 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.561572 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.562379 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.562605 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.562820 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.562983 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.563128 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.563258 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.563468 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564191 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564311 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564398 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564471 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564573 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564654 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564730 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564739 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564807 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.564842 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.573289 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.573411 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.573557 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.573690 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.574971 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5sgbq"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.575387 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-r7t8c"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.575663 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.576213 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.579040 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.581120 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.591801 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.592229 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.592422 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.592596 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.592807 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.592973 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.593702 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.593884 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.594109 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.594297 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rd62r"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.595150 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.598976 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.602052 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.602273 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.602552 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.602945 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.603578 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.604973 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pjtjn"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.605805 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.606335 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.606730 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607008 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.605843 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55b54660-afa7-4a76-b86c-9d1fd3f33769-config\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607315 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b76ts\" (UniqueName: \"kubernetes.io/projected/55b54660-afa7-4a76-b86c-9d1fd3f33769-kube-api-access-b76ts\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607387 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fc9173a7-e308-408a-896b-e2990f50d8c6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w4dg6\" (UID: \"fc9173a7-e308-408a-896b-e2990f50d8c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607454 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55b54660-afa7-4a76-b86c-9d1fd3f33769-serving-cert\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607523 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607552 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-config\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607587 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b240811c-d6e5-4a4f-932a-fdf41ad86a98-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607623 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607650 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55b54660-afa7-4a76-b86c-9d1fd3f33769-trusted-ca\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607696 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-serving-cert\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607737 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b240811c-d6e5-4a4f-932a-fdf41ad86a98-images\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607777 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-audit-policies\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607806 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-encryption-config\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607841 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6877b\" (UniqueName: \"kubernetes.io/projected/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-kube-api-access-6877b\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607874 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqbcf\" (UniqueName: \"kubernetes.io/projected/06f9131b-3fbf-468f-8201-13cc313f7f6c-kube-api-access-wqbcf\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607907 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwltg\" (UniqueName: \"kubernetes.io/projected/2fa6af32-186a-42ac-9513-398cd1124410-kube-api-access-rwltg\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607939 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607966 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/06f9131b-3fbf-468f-8201-13cc313f7f6c-auth-proxy-config\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.607991 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klvpj\" (UniqueName: \"kubernetes.io/projected/fc9173a7-e308-408a-896b-e2990f50d8c6-kube-api-access-klvpj\") pod \"cluster-samples-operator-665b6dd947-w4dg6\" (UID: \"fc9173a7-e308-408a-896b-e2990f50d8c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608023 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-etcd-client\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608052 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f9131b-3fbf-468f-8201-13cc313f7f6c-config\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608074 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-client-ca\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608115 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkx2m\" (UniqueName: \"kubernetes.io/projected/e1a86390-9e3f-4ddf-8440-40b2e4d88412-kube-api-access-rkx2m\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608148 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fa6af32-186a-42ac-9513-398cd1124410-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608182 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608206 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608260 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c2px\" (UniqueName: \"kubernetes.io/projected/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-kube-api-access-7c2px\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608288 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-service-ca-bundle\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608314 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b240811c-d6e5-4a4f-932a-fdf41ad86a98-config\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608343 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-serving-cert\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608414 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6gc2\" (UniqueName: \"kubernetes.io/projected/b240811c-d6e5-4a4f-932a-fdf41ad86a98-kube-api-access-r6gc2\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608455 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-audit-dir\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608481 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-client-ca\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608513 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/06f9131b-3fbf-468f-8201-13cc313f7f6c-machine-approver-tls\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608541 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-config\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608582 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fa6af32-186a-42ac-9513-398cd1124410-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608769 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1a86390-9e3f-4ddf-8440-40b2e4d88412-serving-cert\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608809 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4256fad3-f7d5-4441-a4f0-ee84712d7321-serving-cert\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.608838 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz22k\" (UniqueName: \"kubernetes.io/projected/4256fad3-f7d5-4441-a4f0-ee84712d7321-kube-api-access-zz22k\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.610556 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.613400 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.613597 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.614343 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fb7m9"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.614392 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.614997 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.622255 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.622275 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.622496 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.622598 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.622784 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.622908 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.622925 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.626553 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.634838 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.634932 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.635156 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.636308 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.637968 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fpq7g"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.638311 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.638345 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.638475 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.638649 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.639668 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-68nfv"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.639827 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.640264 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.640818 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.640870 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.644005 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.644931 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.652538 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.653310 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.653321 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.654993 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.658254 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.658902 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.659239 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.659465 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.660125 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fz68t"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.660809 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.660968 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-frs8b"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.661915 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.662239 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.662279 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.663259 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hmx5v"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.664427 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.665276 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.665877 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.666674 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.668679 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.672015 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r7t8c"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.675485 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.677952 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.678136 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.679635 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k5knb"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.680186 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.685852 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.686516 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.687852 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ttqhz"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.693719 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-29qfq"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.697956 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.698792 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.698888 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.699925 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-29qfq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.702650 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tz4gp"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.703678 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.704976 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.706638 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.707059 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.710044 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-78r68"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.710244 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.710921 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.711069 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.711758 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712410 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6877b\" (UniqueName: \"kubernetes.io/projected/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-kube-api-access-6877b\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712455 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqbcf\" (UniqueName: \"kubernetes.io/projected/06f9131b-3fbf-468f-8201-13cc313f7f6c-kube-api-access-wqbcf\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712480 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/533ee8bc-a31e-43a7-b87c-911ecd5c191d-node-pullsecrets\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712499 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwltg\" (UniqueName: \"kubernetes.io/projected/2fa6af32-186a-42ac-9513-398cd1124410-kube-api-access-rwltg\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712518 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-serving-cert\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712813 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-audit-policies\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712852 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82d8b4e3-0fe2-4941-895e-47513a81b9bf-serving-cert\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712860 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5rtlc"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712875 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712903 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712925 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/06f9131b-3fbf-468f-8201-13cc313f7f6c-auth-proxy-config\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712946 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klvpj\" (UniqueName: \"kubernetes.io/projected/fc9173a7-e308-408a-896b-e2990f50d8c6-kube-api-access-klvpj\") pod \"cluster-samples-operator-665b6dd947-w4dg6\" (UID: \"fc9173a7-e308-408a-896b-e2990f50d8c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712965 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.712989 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-metrics-certs\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713007 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/82d8b4e3-0fe2-4941-895e-47513a81b9bf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713025 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/60d29eac-b443-4754-b9bb-d21591d2333d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713044 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713066 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103e81ce-51a3-495a-adf7-87e8af781036-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713088 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-etcd-client\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713106 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f9131b-3fbf-468f-8201-13cc313f7f6c-config\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713126 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-serving-cert\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713146 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-oauth-serving-cert\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713164 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-client-ca\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713185 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60d29eac-b443-4754-b9bb-d21591d2333d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713209 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20f97015-5d3a-43a3-8863-95fcc8dd8414-trusted-ca\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713229 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1b5baf7a-852f-47b1-be0b-3daf981d2281-proxy-tls\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713252 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713272 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-service-ca\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713294 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-config\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713318 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkx2m\" (UniqueName: \"kubernetes.io/projected/e1a86390-9e3f-4ddf-8440-40b2e4d88412-kube-api-access-rkx2m\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713340 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fa6af32-186a-42ac-9513-398cd1124410-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713401 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-client\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713423 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hchw\" (UniqueName: \"kubernetes.io/projected/82d8b4e3-0fe2-4941-895e-47513a81b9bf-kube-api-access-2hchw\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713444 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713464 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713464 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713652 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-config\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713678 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-ca\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713709 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-service-ca-bundle\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713736 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b240811c-d6e5-4a4f-932a-fdf41ad86a98-config\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713779 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c2px\" (UniqueName: \"kubernetes.io/projected/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-kube-api-access-7c2px\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713799 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-etcd-client\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713821 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713845 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20f97015-5d3a-43a3-8863-95fcc8dd8414-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713868 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-serving-cert\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713889 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713912 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-config\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713931 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713952 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713975 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6gc2\" (UniqueName: \"kubernetes.io/projected/b240811c-d6e5-4a4f-932a-fdf41ad86a98-kube-api-access-r6gc2\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.713997 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n7nr\" (UniqueName: \"kubernetes.io/projected/936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8-kube-api-access-5n7nr\") pod \"dns-operator-744455d44c-79tvb\" (UID: \"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8\") " pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714017 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-client-ca\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714037 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d028a5e-d0df-44c7-8e0a-73949ccb3984-config\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714066 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-audit-dir\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714085 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e540236-2888-4509-a53e-9e51cae33546-config\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714103 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e540236-2888-4509-a53e-9e51cae33546-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714126 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80d7820e-ea64-433e-9842-9142a578f408-service-ca-bundle\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714151 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/06f9131b-3fbf-468f-8201-13cc313f7f6c-machine-approver-tls\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714172 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq9s9\" (UniqueName: \"kubernetes.io/projected/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-kube-api-access-qq9s9\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714191 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-stats-auth\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714213 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-config\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714235 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fa6af32-186a-42ac-9513-398cd1124410-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714258 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d028a5e-d0df-44c7-8e0a-73949ccb3984-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714279 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d028a5e-d0df-44c7-8e0a-73949ccb3984-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714303 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714327 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh898\" (UniqueName: \"kubernetes.io/projected/5f3bb363-9319-469e-aca0-29eb769f92ec-kube-api-access-lh898\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714391 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714416 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103e81ce-51a3-495a-adf7-87e8af781036-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714450 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1a86390-9e3f-4ddf-8440-40b2e4d88412-serving-cert\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714478 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvkj7\" (UniqueName: \"kubernetes.io/projected/60d29eac-b443-4754-b9bb-d21591d2333d-kube-api-access-cvkj7\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714500 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8-metrics-tls\") pod \"dns-operator-744455d44c-79tvb\" (UID: \"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8\") " pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714524 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714544 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpbsh\" (UniqueName: \"kubernetes.io/projected/533ee8bc-a31e-43a7-b87c-911ecd5c191d-kube-api-access-rpbsh\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714567 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-default-certificate\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714618 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4256fad3-f7d5-4441-a4f0-ee84712d7321-serving-cert\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714638 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz22k\" (UniqueName: \"kubernetes.io/projected/4256fad3-f7d5-4441-a4f0-ee84712d7321-kube-api-access-zz22k\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714657 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-images\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714677 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skztl\" (UniqueName: \"kubernetes.io/projected/6a9ef22d-4511-403a-9703-fcffd8f5b8e9-kube-api-access-skztl\") pod \"downloads-7954f5f757-r7t8c\" (UID: \"6a9ef22d-4511-403a-9703-fcffd8f5b8e9\") " pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714707 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714727 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55b54660-afa7-4a76-b86c-9d1fd3f33769-config\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714749 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b76ts\" (UniqueName: \"kubernetes.io/projected/55b54660-afa7-4a76-b86c-9d1fd3f33769-kube-api-access-b76ts\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714769 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fc9173a7-e308-408a-896b-e2990f50d8c6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w4dg6\" (UID: \"fc9173a7-e308-408a-896b-e2990f50d8c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714789 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-etcd-serving-ca\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714812 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-service-ca\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714834 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714870 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55b54660-afa7-4a76-b86c-9d1fd3f33769-serving-cert\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714893 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn5h4\" (UniqueName: \"kubernetes.io/projected/80d7820e-ea64-433e-9842-9142a578f408-kube-api-access-cn5h4\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714914 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-audit\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714936 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-encryption-config\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714963 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.714983 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-proxy-tls\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715006 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1b5baf7a-852f-47b1-be0b-3daf981d2281-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715004 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-service-ca-bundle\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715025 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjmg9\" (UniqueName: \"kubernetes.io/projected/8e339a50-fbf4-44da-9335-ba6ae033fc38-kube-api-access-hjmg9\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715047 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e540236-2888-4509-a53e-9e51cae33546-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715067 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/103e81ce-51a3-495a-adf7-87e8af781036-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715091 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-config\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715113 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzjhj\" (UniqueName: \"kubernetes.io/projected/20f97015-5d3a-43a3-8863-95fcc8dd8414-kube-api-access-zzjhj\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715134 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-image-import-ca\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715124 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715152 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-trusted-ca-bundle\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715251 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715280 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715308 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd501849-1d14-48c1-bc51-06d23c56db28-audit-dir\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715340 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b240811c-d6e5-4a4f-932a-fdf41ad86a98-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715381 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20f97015-5d3a-43a3-8863-95fcc8dd8414-metrics-tls\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715402 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-oauth-config\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715426 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wzdt\" (UniqueName: \"kubernetes.io/projected/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-kube-api-access-5wzdt\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715451 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttbdv\" (UniqueName: \"kubernetes.io/projected/1b5baf7a-852f-47b1-be0b-3daf981d2281-kube-api-access-ttbdv\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715477 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715503 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55b54660-afa7-4a76-b86c-9d1fd3f33769-trusted-ca\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715526 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f3bb363-9319-469e-aca0-29eb769f92ec-serving-cert\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715546 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hmmk\" (UniqueName: \"kubernetes.io/projected/dd501849-1d14-48c1-bc51-06d23c56db28-kube-api-access-9hmmk\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715579 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-serving-cert\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715599 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b240811c-d6e5-4a4f-932a-fdf41ad86a98-images\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715618 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/60d29eac-b443-4754-b9bb-d21591d2333d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715640 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-audit-policies\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715661 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-encryption-config\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.715690 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/533ee8bc-a31e-43a7-b87c-911ecd5c191d-audit-dir\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.716312 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.716774 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fa6af32-186a-42ac-9513-398cd1124410-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.717164 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.717206 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/06f9131b-3fbf-468f-8201-13cc313f7f6c-auth-proxy-config\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.717378 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06f9131b-3fbf-468f-8201-13cc313f7f6c-config\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.717913 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b240811c-d6e5-4a4f-932a-fdf41ad86a98-images\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.717949 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55b54660-afa7-4a76-b86c-9d1fd3f33769-trusted-ca\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.718538 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-blzx5"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.721676 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.722021 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.722383 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.723082 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-audit-policies\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.723351 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.723624 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-audit-dir\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.725776 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rd62r"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.726605 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-encryption-config\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.727738 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.727910 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55b54660-afa7-4a76-b86c-9d1fd3f33769-serving-cert\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.727933 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-etcd-client\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.730021 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.730228 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b240811c-d6e5-4a4f-932a-fdf41ad86a98-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.730707 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-serving-cert\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.730971 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/06f9131b-3fbf-468f-8201-13cc313f7f6c-machine-approver-tls\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.731014 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1a86390-9e3f-4ddf-8440-40b2e4d88412-config\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.731767 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-79tvb"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.733182 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.734537 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.736104 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fpq7g"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.736131 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5sgbq"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.738890 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-frs8b"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.738952 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.739543 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.741464 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fa6af32-186a-42ac-9513-398cd1124410-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.741639 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pbx8v"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.741676 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fz68t"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.742546 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pjtjn"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.748070 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.750595 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.757151 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-blzx5"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.757177 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.758113 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.760035 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.763224 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.764707 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.765954 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-29qfq"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.767244 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.768611 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cxgpw"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.769323 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4256fad3-f7d5-4441-a4f0-ee84712d7321-serving-cert\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.769327 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-serving-cert\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.769908 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.770127 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55b54660-afa7-4a76-b86c-9d1fd3f33769-config\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.771123 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k5knb"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.771488 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1a86390-9e3f-4ddf-8440-40b2e4d88412-serving-cert\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.771525 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-client-ca\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.771691 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-client-ca\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.772013 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-config\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.775150 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/fc9173a7-e308-408a-896b-e2990f50d8c6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-w4dg6\" (UID: \"fc9173a7-e308-408a-896b-e2990f50d8c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.776594 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.777057 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.779237 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-78r68"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.780493 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b240811c-d6e5-4a4f-932a-fdf41ad86a98-config\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.780767 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.782899 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ttqhz"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.786659 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tz4gp"] Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.797415 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.816521 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttbdv\" (UniqueName: \"kubernetes.io/projected/1b5baf7a-852f-47b1-be0b-3daf981d2281-kube-api-access-ttbdv\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.816571 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-certs\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.816598 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-apiservice-cert\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.816632 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/533ee8bc-a31e-43a7-b87c-911ecd5c191d-audit-dir\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.816669 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-serving-cert\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.816692 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-audit-policies\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.816727 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817321 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817436 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817452 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/533ee8bc-a31e-43a7-b87c-911ecd5c191d-audit-dir\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817481 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/82d8b4e3-0fe2-4941-895e-47513a81b9bf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817623 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/60d29eac-b443-4754-b9bb-d21591d2333d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817658 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-serving-cert\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817683 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-srv-cert\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817708 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20f97015-5d3a-43a3-8863-95fcc8dd8414-trusted-ca\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817732 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-service-ca\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817752 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/aee1200b-878a-4a90-a34c-dcc237c6a8a3-tmpfs\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817774 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmcph\" (UniqueName: \"kubernetes.io/projected/35350b4a-c3a6-4d70-bab2-a79757701029-kube-api-access-jmcph\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817798 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-config\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817835 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35350b4a-c3a6-4d70-bab2-a79757701029-config\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817860 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hchw\" (UniqueName: \"kubernetes.io/projected/82d8b4e3-0fe2-4941-895e-47513a81b9bf-kube-api-access-2hchw\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817868 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/82d8b4e3-0fe2-4941-895e-47513a81b9bf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817883 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-ca\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817907 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817931 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20f97015-5d3a-43a3-8863-95fcc8dd8414-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817956 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35350b4a-c3a6-4d70-bab2-a79757701029-serving-cert\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817972 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818078 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-audit-policies\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.817979 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-config\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818173 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818220 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7nr\" (UniqueName: \"kubernetes.io/projected/936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8-kube-api-access-5n7nr\") pod \"dns-operator-744455d44c-79tvb\" (UID: \"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8\") " pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818246 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq9s9\" (UniqueName: \"kubernetes.io/projected/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-kube-api-access-qq9s9\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818265 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rksp5\" (UniqueName: \"kubernetes.io/projected/62b435a3-f5a3-4cef-a80c-5968f148d816-kube-api-access-rksp5\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818284 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d028a5e-d0df-44c7-8e0a-73949ccb3984-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818300 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh898\" (UniqueName: \"kubernetes.io/projected/5f3bb363-9319-469e-aca0-29eb769f92ec-kube-api-access-lh898\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818317 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818333 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn8zx\" (UniqueName: \"kubernetes.io/projected/aee1200b-878a-4a90-a34c-dcc237c6a8a3-kube-api-access-dn8zx\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818349 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-default-certificate\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818393 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818412 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpbsh\" (UniqueName: \"kubernetes.io/projected/533ee8bc-a31e-43a7-b87c-911ecd5c191d-kube-api-access-rpbsh\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818430 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-images\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818446 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-profile-collector-cert\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818464 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-etcd-serving-ca\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818480 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-service-ca\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818497 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818516 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818569 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjmg9\" (UniqueName: \"kubernetes.io/projected/8e339a50-fbf4-44da-9335-ba6ae033fc38-kube-api-access-hjmg9\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818586 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/103e81ce-51a3-495a-adf7-87e8af781036-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818611 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-proxy-tls\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818628 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1b5baf7a-852f-47b1-be0b-3daf981d2281-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818632 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-config\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818644 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-audit\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818660 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-image-import-ca\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818675 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-trusted-ca-bundle\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818695 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20f97015-5d3a-43a3-8863-95fcc8dd8414-metrics-tls\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818713 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wzdt\" (UniqueName: \"kubernetes.io/projected/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-kube-api-access-5wzdt\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818730 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f3bb363-9319-469e-aca0-29eb769f92ec-serving-cert\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818748 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6p97\" (UniqueName: \"kubernetes.io/projected/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-kube-api-access-t6p97\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818766 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/60d29eac-b443-4754-b9bb-d21591d2333d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818782 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hmmk\" (UniqueName: \"kubernetes.io/projected/dd501849-1d14-48c1-bc51-06d23c56db28-kube-api-access-9hmmk\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818796 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818817 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/533ee8bc-a31e-43a7-b87c-911ecd5c191d-node-pullsecrets\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818833 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818855 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82d8b4e3-0fe2-4941-895e-47513a81b9bf-serving-cert\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818871 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-metrics-certs\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818887 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-webhook-cert\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818904 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818919 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103e81ce-51a3-495a-adf7-87e8af781036-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818935 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-oauth-serving-cert\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818951 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60d29eac-b443-4754-b9bb-d21591d2333d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818967 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1b5baf7a-852f-47b1-be0b-3daf981d2281-proxy-tls\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.818984 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819007 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-node-bootstrap-token\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819023 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-client\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819041 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-config\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819057 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-etcd-client\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819108 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819131 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n87fq\" (UniqueName: \"kubernetes.io/projected/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-kube-api-access-n87fq\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819155 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819185 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d028a5e-d0df-44c7-8e0a-73949ccb3984-config\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819208 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e540236-2888-4509-a53e-9e51cae33546-config\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819228 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e540236-2888-4509-a53e-9e51cae33546-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819249 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80d7820e-ea64-433e-9842-9142a578f408-service-ca-bundle\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819269 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-stats-auth\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819292 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d028a5e-d0df-44c7-8e0a-73949ccb3984-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819314 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819336 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvkj7\" (UniqueName: \"kubernetes.io/projected/60d29eac-b443-4754-b9bb-d21591d2333d-kube-api-access-cvkj7\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819389 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8-metrics-tls\") pod \"dns-operator-744455d44c-79tvb\" (UID: \"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8\") " pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819415 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103e81ce-51a3-495a-adf7-87e8af781036-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819459 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skztl\" (UniqueName: \"kubernetes.io/projected/6a9ef22d-4511-403a-9703-fcffd8f5b8e9-kube-api-access-skztl\") pod \"downloads-7954f5f757-r7t8c\" (UID: \"6a9ef22d-4511-403a-9703-fcffd8f5b8e9\") " pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819504 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn5h4\" (UniqueName: \"kubernetes.io/projected/80d7820e-ea64-433e-9842-9142a578f408-kube-api-access-cn5h4\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819528 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e540236-2888-4509-a53e-9e51cae33546-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819553 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-encryption-config\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819576 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzjhj\" (UniqueName: \"kubernetes.io/projected/20f97015-5d3a-43a3-8863-95fcc8dd8414-kube-api-access-zzjhj\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819601 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819625 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819639 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819647 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-oauth-config\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819669 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd501849-1d14-48c1-bc51-06d23c56db28-audit-dir\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819726 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-serving-cert\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.819753 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd501849-1d14-48c1-bc51-06d23c56db28-audit-dir\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.820431 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.820544 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.820922 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/533ee8bc-a31e-43a7-b87c-911ecd5c191d-node-pullsecrets\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.821020 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/60d29eac-b443-4754-b9bb-d21591d2333d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.821085 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-audit\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.821150 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.821576 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1b5baf7a-852f-47b1-be0b-3daf981d2281-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.821655 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e540236-2888-4509-a53e-9e51cae33546-config\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.821752 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-etcd-serving-ca\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.821813 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-service-ca\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.822601 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.822722 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-config\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.824094 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-oauth-serving-cert\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.824202 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60d29eac-b443-4754-b9bb-d21591d2333d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.824245 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.824571 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.824680 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-trusted-ca-bundle\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.825515 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.826069 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-encryption-config\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.826143 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.826620 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.826757 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/533ee8bc-a31e-43a7-b87c-911ecd5c191d-etcd-client\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.826813 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82d8b4e3-0fe2-4941-895e-47513a81b9bf-serving-cert\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.827102 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.827203 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.827206 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-serving-cert\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.827291 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e540236-2888-4509-a53e-9e51cae33546-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.828678 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.829137 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-oauth-config\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.829384 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8-metrics-tls\") pod \"dns-operator-744455d44c-79tvb\" (UID: \"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8\") " pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.832582 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.838241 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.843444 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/533ee8bc-a31e-43a7-b87c-911ecd5c191d-image-import-ca\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.857343 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.876968 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.897951 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.906131 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/103e81ce-51a3-495a-adf7-87e8af781036-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.916485 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.920351 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-node-bootstrap-token\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.920505 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n87fq\" (UniqueName: \"kubernetes.io/projected/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-kube-api-access-n87fq\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.920726 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-certs\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.920841 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-apiservice-cert\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.920976 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-srv-cert\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921143 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/aee1200b-878a-4a90-a34c-dcc237c6a8a3-tmpfs\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921228 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmcph\" (UniqueName: \"kubernetes.io/projected/35350b4a-c3a6-4d70-bab2-a79757701029-kube-api-access-jmcph\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921302 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35350b4a-c3a6-4d70-bab2-a79757701029-config\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921428 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35350b4a-c3a6-4d70-bab2-a79757701029-serving-cert\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921539 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rksp5\" (UniqueName: \"kubernetes.io/projected/62b435a3-f5a3-4cef-a80c-5968f148d816-kube-api-access-rksp5\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921620 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn8zx\" (UniqueName: \"kubernetes.io/projected/aee1200b-878a-4a90-a34c-dcc237c6a8a3-kube-api-access-dn8zx\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921718 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-profile-collector-cert\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921833 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6p97\" (UniqueName: \"kubernetes.io/projected/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-kube-api-access-t6p97\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921921 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921996 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.921724 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/aee1200b-878a-4a90-a34c-dcc237c6a8a3-tmpfs\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.922080 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-webhook-cert\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.922292 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/103e81ce-51a3-495a-adf7-87e8af781036-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.956998 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.966099 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-client\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.976836 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 04:35:18 crc kubenswrapper[4992]: I1004 04:35:18.997154 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.000348 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-config\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.016652 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.033270 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f3bb363-9319-469e-aca0-29eb769f92ec-serving-cert\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.038099 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.041669 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-service-ca\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.058028 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.076805 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.077728 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5f3bb363-9319-469e-aca0-29eb769f92ec-etcd-ca\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.096935 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.116941 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.126701 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1b5baf7a-852f-47b1-be0b-3daf981d2281-proxy-tls\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.137633 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.158985 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.163477 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20f97015-5d3a-43a3-8863-95fcc8dd8414-metrics-tls\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.177791 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.202893 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.211809 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20f97015-5d3a-43a3-8863-95fcc8dd8414-trusted-ca\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.217689 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.237589 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.257382 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.277328 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.287589 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-stats-auth\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.297174 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.302142 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d028a5e-d0df-44c7-8e0a-73949ccb3984-config\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.317624 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.337663 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.357692 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.377990 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.388535 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-metrics-certs\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.398058 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.404945 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80d7820e-ea64-433e-9842-9142a578f408-service-ca-bundle\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.417145 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.424678 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/80d7820e-ea64-433e-9842-9142a578f408-default-certificate\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.437630 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.457874 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.471026 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d028a5e-d0df-44c7-8e0a-73949ccb3984-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.477125 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.481486 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-images\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.498076 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.518137 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.526338 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-proxy-tls\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.538573 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.557867 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.578029 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.583267 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.598999 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.605336 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.617013 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.637246 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.656829 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.676142 4992 request.go:700] Waited for 1.016386934s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.678162 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.698387 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.717426 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.722794 4992 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.722874 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config podName:4256fad3-f7d5-4441-a4f0-ee84712d7321 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:20.222856454 +0000 UTC m=+154.070531922 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config") pod "route-controller-manager-6576b87f9c-4tqf8" (UID: "4256fad3-f7d5-4441-a4f0-ee84712d7321") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.737025 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.758219 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.778240 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.797732 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.818487 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.837703 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.857863 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.865819 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-profile-collector-cert\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.879638 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.898905 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.917151 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.920827 4992 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.920937 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-node-bootstrap-token podName:62b435a3-f5a3-4cef-a80c-5968f148d816 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:20.420887745 +0000 UTC m=+154.268563213 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-node-bootstrap-token") pod "machine-config-server-5rtlc" (UID: "62b435a3-f5a3-4cef-a80c-5968f148d816") : failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.921213 4992 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.921266 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-srv-cert podName:86a9fb33-b1f8-4f50-9785-a5504e4b4efd nodeName:}" failed. No retries permitted until 2025-10-04 04:35:20.421249455 +0000 UTC m=+154.268924933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-srv-cert") pod "catalog-operator-68c6474976-rplg6" (UID: "86a9fb33-b1f8-4f50-9785-a5504e4b4efd") : failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.921297 4992 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.921328 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-apiservice-cert podName:aee1200b-878a-4a90-a34c-dcc237c6a8a3 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:20.421320607 +0000 UTC m=+154.268996075 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-apiservice-cert") pod "packageserver-d55dfcdfc-xjmbw" (UID: "aee1200b-878a-4a90-a34c-dcc237c6a8a3") : failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.921353 4992 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.921437 4992 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.921471 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-certs podName:62b435a3-f5a3-4cef-a80c-5968f148d816 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:20.42144653 +0000 UTC m=+154.269121998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-certs") pod "machine-config-server-5rtlc" (UID: "62b435a3-f5a3-4cef-a80c-5968f148d816") : failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.921522 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35350b4a-c3a6-4d70-bab2-a79757701029-config podName:35350b4a-c3a6-4d70-bab2-a79757701029 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:20.421480221 +0000 UTC m=+154.269155899 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/35350b4a-c3a6-4d70-bab2-a79757701029-config") pod "service-ca-operator-777779d784-k5knb" (UID: "35350b4a-c3a6-4d70-bab2-a79757701029") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.922555 4992 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.922653 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35350b4a-c3a6-4d70-bab2-a79757701029-serving-cert podName:35350b4a-c3a6-4d70-bab2-a79757701029 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:20.422632862 +0000 UTC m=+154.270308330 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/35350b4a-c3a6-4d70-bab2-a79757701029-serving-cert") pod "service-ca-operator-777779d784-k5knb" (UID: "35350b4a-c3a6-4d70-bab2-a79757701029") : failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.922567 4992 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: E1004 04:35:19.922705 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-webhook-cert podName:aee1200b-878a-4a90-a34c-dcc237c6a8a3 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:20.422697113 +0000 UTC m=+154.270372581 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-webhook-cert") pod "packageserver-d55dfcdfc-xjmbw" (UID: "aee1200b-878a-4a90-a34c-dcc237c6a8a3") : failed to sync secret cache: timed out waiting for the condition Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.936514 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.958019 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.978238 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 04:35:19 crc kubenswrapper[4992]: I1004 04:35:19.998433 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.017281 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.037111 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.057277 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.078330 4992 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.097871 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.116821 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.137002 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.157290 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.176700 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.197712 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.217413 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.237110 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.241707 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.257319 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.285966 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.297230 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.316569 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.336783 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.357516 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.377513 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.396849 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.417831 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.438019 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.445544 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-certs\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.445626 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-apiservice-cert\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.445701 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-srv-cert\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.445752 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35350b4a-c3a6-4d70-bab2-a79757701029-config\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.445848 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35350b4a-c3a6-4d70-bab2-a79757701029-serving-cert\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.446089 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-webhook-cert\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.446191 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-node-bootstrap-token\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.446904 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35350b4a-c3a6-4d70-bab2-a79757701029-config\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.448979 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-apiservice-cert\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.449886 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35350b4a-c3a6-4d70-bab2-a79757701029-serving-cert\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.451541 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-srv-cert\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.451666 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aee1200b-878a-4a90-a34c-dcc237c6a8a3-webhook-cert\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.457962 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.496049 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqbcf\" (UniqueName: \"kubernetes.io/projected/06f9131b-3fbf-468f-8201-13cc313f7f6c-kube-api-access-wqbcf\") pod \"machine-approver-56656f9798-h86m4\" (UID: \"06f9131b-3fbf-468f-8201-13cc313f7f6c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.513848 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6877b\" (UniqueName: \"kubernetes.io/projected/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-kube-api-access-6877b\") pod \"controller-manager-879f6c89f-fb7m9\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.531511 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwltg\" (UniqueName: \"kubernetes.io/projected/2fa6af32-186a-42ac-9513-398cd1124410-kube-api-access-rwltg\") pod \"openshift-apiserver-operator-796bbdcf4f-tbcnm\" (UID: \"2fa6af32-186a-42ac-9513-398cd1124410\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.536307 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.549527 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-certs\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.575195 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c2px\" (UniqueName: \"kubernetes.io/projected/3cd8a6e8-ee96-4e7e-9367-e74478678f5f-kube-api-access-7c2px\") pod \"apiserver-7bbb656c7d-m8ccp\" (UID: \"3cd8a6e8-ee96-4e7e-9367-e74478678f5f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.577319 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.597441 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.613074 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/62b435a3-f5a3-4cef-a80c-5968f148d816-node-bootstrap-token\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.632264 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkx2m\" (UniqueName: \"kubernetes.io/projected/e1a86390-9e3f-4ddf-8440-40b2e4d88412-kube-api-access-rkx2m\") pod \"authentication-operator-69f744f599-cxgpw\" (UID: \"e1a86390-9e3f-4ddf-8440-40b2e4d88412\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.645021 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.654943 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klvpj\" (UniqueName: \"kubernetes.io/projected/fc9173a7-e308-408a-896b-e2990f50d8c6-kube-api-access-klvpj\") pod \"cluster-samples-operator-665b6dd947-w4dg6\" (UID: \"fc9173a7-e308-408a-896b-e2990f50d8c6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" Oct 04 04:35:20 crc kubenswrapper[4992]: W1004 04:35:20.657023 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06f9131b_3fbf_468f_8201_13cc313f7f6c.slice/crio-c5543e0217e4edda4acebde7a833f0841cda2edecc4d3ead56cbbf40a1d6d021 WatchSource:0}: Error finding container c5543e0217e4edda4acebde7a833f0841cda2edecc4d3ead56cbbf40a1d6d021: Status 404 returned error can't find the container with id c5543e0217e4edda4acebde7a833f0841cda2edecc4d3ead56cbbf40a1d6d021 Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.658714 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.675196 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6gc2\" (UniqueName: \"kubernetes.io/projected/b240811c-d6e5-4a4f-932a-fdf41ad86a98-kube-api-access-r6gc2\") pod \"machine-api-operator-5694c8668f-pbx8v\" (UID: \"b240811c-d6e5-4a4f-932a-fdf41ad86a98\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.678710 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.691731 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.695417 4992 request.go:700] Waited for 1.972586679s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.697408 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.717707 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.734139 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.760516 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.761829 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b76ts\" (UniqueName: \"kubernetes.io/projected/55b54660-afa7-4a76-b86c-9d1fd3f33769-kube-api-access-b76ts\") pod \"console-operator-58897d9998-hmx5v\" (UID: \"55b54660-afa7-4a76-b86c-9d1fd3f33769\") " pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.766564 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.778150 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz22k\" (UniqueName: \"kubernetes.io/projected/4256fad3-f7d5-4441-a4f0-ee84712d7321-kube-api-access-zz22k\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.787807 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.815441 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttbdv\" (UniqueName: \"kubernetes.io/projected/1b5baf7a-852f-47b1-be0b-3daf981d2281-kube-api-access-ttbdv\") pod \"machine-config-controller-84d6567774-9kblj\" (UID: \"1b5baf7a-852f-47b1-be0b-3daf981d2281\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.836102 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hchw\" (UniqueName: \"kubernetes.io/projected/82d8b4e3-0fe2-4941-895e-47513a81b9bf-kube-api-access-2hchw\") pod \"openshift-config-operator-7777fb866f-ftjbd\" (UID: \"82d8b4e3-0fe2-4941-895e-47513a81b9bf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.854944 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20f97015-5d3a-43a3-8863-95fcc8dd8414-bound-sa-token\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.872611 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n7nr\" (UniqueName: \"kubernetes.io/projected/936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8-kube-api-access-5n7nr\") pod \"dns-operator-744455d44c-79tvb\" (UID: \"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8\") " pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.891886 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq9s9\" (UniqueName: \"kubernetes.io/projected/c2c1be78-1a57-47ca-a671-cb0a43e05c2a-kube-api-access-qq9s9\") pod \"openshift-controller-manager-operator-756b6f6bc6-gh7xf\" (UID: \"c2c1be78-1a57-47ca-a671-cb0a43e05c2a\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.896124 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.896191 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.906958 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fb7m9"] Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.918836 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp"] Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.920230 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d028a5e-d0df-44c7-8e0a-73949ccb3984-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8zxrr\" (UID: \"3d028a5e-d0df-44c7-8e0a-73949ccb3984\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.932679 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh898\" (UniqueName: \"kubernetes.io/projected/5f3bb363-9319-469e-aca0-29eb769f92ec-kube-api-access-lh898\") pod \"etcd-operator-b45778765-fpq7g\" (UID: \"5f3bb363-9319-469e-aca0-29eb769f92ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.942882 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.949459 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.962913 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-pbx8v"] Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.963088 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.963194 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/103e81ce-51a3-495a-adf7-87e8af781036-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n9lrz\" (UID: \"103e81ce-51a3-495a-adf7-87e8af781036\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.979901 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjmg9\" (UniqueName: \"kubernetes.io/projected/8e339a50-fbf4-44da-9335-ba6ae033fc38-kube-api-access-hjmg9\") pod \"console-f9d7485db-5sgbq\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:20 crc kubenswrapper[4992]: I1004 04:35:20.997114 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpbsh\" (UniqueName: \"kubernetes.io/projected/533ee8bc-a31e-43a7-b87c-911ecd5c191d-kube-api-access-rpbsh\") pod \"apiserver-76f77b778f-pjtjn\" (UID: \"533ee8bc-a31e-43a7-b87c-911ecd5c191d\") " pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.018225 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/60d29eac-b443-4754-b9bb-d21591d2333d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.029547 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" event={"ID":"d8c6ed95-6bcb-4a2d-b7d4-de1419159770","Type":"ContainerStarted","Data":"f98ab7d9a4d88da5bd4ebf6e77dc6435ff8659e2e3c10efc6c482ec64b6bfd76"} Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.031063 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" event={"ID":"b240811c-d6e5-4a4f-932a-fdf41ad86a98","Type":"ContainerStarted","Data":"61e64c4462ec563224f9a5f1133ead775cd313fa8280434634c0283de72da959"} Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.031745 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.035584 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" event={"ID":"3cd8a6e8-ee96-4e7e-9367-e74478678f5f","Type":"ContainerStarted","Data":"6e42159b4c94c799982d04ea0e355af11bc4a1b039b049d42c1ef5cb0b1c5ed7"} Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.038459 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" event={"ID":"06f9131b-3fbf-468f-8201-13cc313f7f6c","Type":"ContainerStarted","Data":"c5543e0217e4edda4acebde7a833f0841cda2edecc4d3ead56cbbf40a1d6d021"} Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.039047 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wzdt\" (UniqueName: \"kubernetes.io/projected/78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d-kube-api-access-5wzdt\") pod \"machine-config-operator-74547568cd-fwb7s\" (UID: \"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.087546 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvkj7\" (UniqueName: \"kubernetes.io/projected/60d29eac-b443-4754-b9bb-d21591d2333d-kube-api-access-cvkj7\") pod \"cluster-image-registry-operator-dc59b4c8b-xrgdh\" (UID: \"60d29eac-b443-4754-b9bb-d21591d2333d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.101308 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.107369 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hmmk\" (UniqueName: \"kubernetes.io/projected/dd501849-1d14-48c1-bc51-06d23c56db28-kube-api-access-9hmmk\") pod \"oauth-openshift-558db77b4-rd62r\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.113912 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skztl\" (UniqueName: \"kubernetes.io/projected/6a9ef22d-4511-403a-9703-fcffd8f5b8e9-kube-api-access-skztl\") pod \"downloads-7954f5f757-r7t8c\" (UID: \"6a9ef22d-4511-403a-9703-fcffd8f5b8e9\") " pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.115778 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.126139 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e540236-2888-4509-a53e-9e51cae33546-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-b79wf\" (UID: \"9e540236-2888-4509-a53e-9e51cae33546\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.136436 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm"] Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.159195 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzjhj\" (UniqueName: \"kubernetes.io/projected/20f97015-5d3a-43a3-8863-95fcc8dd8414-kube-api-access-zzjhj\") pod \"ingress-operator-5b745b69d9-b99bg\" (UID: \"20f97015-5d3a-43a3-8863-95fcc8dd8414\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:21 crc kubenswrapper[4992]: W1004 04:35:21.164331 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fa6af32_186a_42ac_9513_398cd1124410.slice/crio-80a93eed1d902d287072af68f80d072b57cca2652bfd51682f7be3bd389efa66 WatchSource:0}: Error finding container 80a93eed1d902d287072af68f80d072b57cca2652bfd51682f7be3bd389efa66: Status 404 returned error can't find the container with id 80a93eed1d902d287072af68f80d072b57cca2652bfd51682f7be3bd389efa66 Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.170819 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.172184 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.174848 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.176648 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n87fq\" (UniqueName: \"kubernetes.io/projected/86a9fb33-b1f8-4f50-9785-a5504e4b4efd-kube-api-access-n87fq\") pod \"catalog-operator-68c6474976-rplg6\" (UID: \"86a9fb33-b1f8-4f50-9785-a5504e4b4efd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.176947 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn5h4\" (UniqueName: \"kubernetes.io/projected/80d7820e-ea64-433e-9842-9142a578f408-kube-api-access-cn5h4\") pod \"router-default-5444994796-68nfv\" (UID: \"80d7820e-ea64-433e-9842-9142a578f408\") " pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.183133 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.191510 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.200753 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.201828 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmcph\" (UniqueName: \"kubernetes.io/projected/35350b4a-c3a6-4d70-bab2-a79757701029-kube-api-access-jmcph\") pod \"service-ca-operator-777779d784-k5knb\" (UID: \"35350b4a-c3a6-4d70-bab2-a79757701029\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.207867 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.213767 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cxgpw"] Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.214411 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.222153 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rksp5\" (UniqueName: \"kubernetes.io/projected/62b435a3-f5a3-4cef-a80c-5968f148d816-kube-api-access-rksp5\") pod \"machine-config-server-5rtlc\" (UID: \"62b435a3-f5a3-4cef-a80c-5968f148d816\") " pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.242230 4992 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.242394 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config podName:4256fad3-f7d5-4441-a4f0-ee84712d7321 nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.242342404 +0000 UTC m=+156.090017872 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config") pod "route-controller-manager-6576b87f9c-4tqf8" (UID: "4256fad3-f7d5-4441-a4f0-ee84712d7321") : failed to sync configmap cache: timed out waiting for the condition Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.245962 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn8zx\" (UniqueName: \"kubernetes.io/projected/aee1200b-878a-4a90-a34c-dcc237c6a8a3-kube-api-access-dn8zx\") pod \"packageserver-d55dfcdfc-xjmbw\" (UID: \"aee1200b-878a-4a90-a34c-dcc237c6a8a3\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.261168 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.280521 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.282313 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6p97\" (UniqueName: \"kubernetes.io/projected/f796a803-cdb3-46bf-82f0-9e89f7f50a4a-kube-api-access-t6p97\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cq9q\" (UID: \"f796a803-cdb3-46bf-82f0-9e89f7f50a4a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.283123 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.283290 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.284529 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.294802 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6"] Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.325923 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.334694 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.366168 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.371673 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmjpk\" (UniqueName: \"kubernetes.io/projected/617d3d62-5b4e-481b-817c-54288406fc7c-kube-api-access-dmjpk\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.371835 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-trusted-ca\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.371866 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glx8g\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-kube-api-access-glx8g\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.371891 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0784a997-5d78-4199-a3fa-3360eef7a756-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8c9n2\" (UID: \"0784a997-5d78-4199-a3fa-3360eef7a756\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.373806 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-mountpoint-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.373944 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-registration-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.373997 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-metrics-tls\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374067 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374205 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fssv5\" (UniqueName: \"kubernetes.io/projected/9d143445-9a26-4203-995d-82238aa12f54-kube-api-access-fssv5\") pod \"control-plane-machine-set-operator-78cbb6b69f-4f9mb\" (UID: \"9d143445-9a26-4203-995d-82238aa12f54\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374235 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-socket-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374411 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc24k\" (UniqueName: \"kubernetes.io/projected/26b1f854-1ae7-4987-9454-94263ed6b5d7-kube-api-access-fc24k\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374501 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/26b1f854-1ae7-4987-9454-94263ed6b5d7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374551 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-signing-cabundle\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374674 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b209aa10-3135-455f-8ad0-a54f64b7768d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-frs8b\" (UID: \"b209aa10-3135-455f-8ad0-a54f64b7768d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374705 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/26b1f854-1ae7-4987-9454-94263ed6b5d7-srv-cert\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374791 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78372347-a30d-4a0f-a4d4-c92a1999eaad-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374841 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pr2h\" (UniqueName: \"kubernetes.io/projected/0b337efa-c799-4fce-833c-cd0149c3644c-kube-api-access-9pr2h\") pod \"migrator-59844c95c7-x8sls\" (UID: \"0b337efa-c799-4fce-833c-cd0149c3644c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.374983 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm5lq\" (UniqueName: \"kubernetes.io/projected/24abe58b-4b74-4207-ba8f-1b16c22617cc-kube-api-access-bm5lq\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.375011 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-plugins-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.375107 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8v4q\" (UniqueName: \"kubernetes.io/projected/5b11a69d-9e3a-4e77-8c17-a9f2cc34873d-kube-api-access-k8v4q\") pod \"ingress-canary-29qfq\" (UID: \"5b11a69d-9e3a-4e77-8c17-a9f2cc34873d\") " pod="openshift-ingress-canary/ingress-canary-29qfq" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.375189 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.375240 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjr97\" (UniqueName: \"kubernetes.io/projected/ee4b8c43-5789-4d06-884e-f2fcc355eeec-kube-api-access-hjr97\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.376097 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-bound-sa-token\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.376220 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59clc\" (UniqueName: \"kubernetes.io/projected/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-kube-api-access-59clc\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377376 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d143445-9a26-4203-995d-82238aa12f54-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4f9mb\" (UID: \"9d143445-9a26-4203-995d-82238aa12f54\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377451 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gftn\" (UniqueName: \"kubernetes.io/projected/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-kube-api-access-8gftn\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377504 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78372347-a30d-4a0f-a4d4-c92a1999eaad-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377535 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24abe58b-4b74-4207-ba8f-1b16c22617cc-config-volume\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377584 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nln87\" (UniqueName: \"kubernetes.io/projected/b209aa10-3135-455f-8ad0-a54f64b7768d-kube-api-access-nln87\") pod \"multus-admission-controller-857f4d67dd-frs8b\" (UID: \"b209aa10-3135-455f-8ad0-a54f64b7768d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377610 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vv9n\" (UniqueName: \"kubernetes.io/projected/0784a997-5d78-4199-a3fa-3360eef7a756-kube-api-access-2vv9n\") pod \"package-server-manager-789f6589d5-8c9n2\" (UID: \"0784a997-5d78-4199-a3fa-3360eef7a756\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377670 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24abe58b-4b74-4207-ba8f-1b16c22617cc-secret-volume\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377715 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-signing-key\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.377944 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.378052 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-csi-data-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.378095 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-config-volume\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.378246 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-tls\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.378307 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b11a69d-9e3a-4e77-8c17-a9f2cc34873d-cert\") pod \"ingress-canary-29qfq\" (UID: \"5b11a69d-9e3a-4e77-8c17-a9f2cc34873d\") " pod="openshift-ingress-canary/ingress-canary-29qfq" Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.378866 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:21.878848311 +0000 UTC m=+155.726523779 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.381178 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-certificates\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.410388 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5rtlc" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.484569 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.484792 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:21.984747324 +0000 UTC m=+155.832422792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485308 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0784a997-5d78-4199-a3fa-3360eef7a756-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8c9n2\" (UID: \"0784a997-5d78-4199-a3fa-3360eef7a756\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485345 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-mountpoint-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485449 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-registration-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485492 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-metrics-tls\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485564 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485596 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fssv5\" (UniqueName: \"kubernetes.io/projected/9d143445-9a26-4203-995d-82238aa12f54-kube-api-access-fssv5\") pod \"control-plane-machine-set-operator-78cbb6b69f-4f9mb\" (UID: \"9d143445-9a26-4203-995d-82238aa12f54\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485614 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-socket-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485645 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc24k\" (UniqueName: \"kubernetes.io/projected/26b1f854-1ae7-4987-9454-94263ed6b5d7-kube-api-access-fc24k\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485672 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/26b1f854-1ae7-4987-9454-94263ed6b5d7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485691 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-signing-cabundle\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485722 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b209aa10-3135-455f-8ad0-a54f64b7768d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-frs8b\" (UID: \"b209aa10-3135-455f-8ad0-a54f64b7768d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485737 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/26b1f854-1ae7-4987-9454-94263ed6b5d7-srv-cert\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485757 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78372347-a30d-4a0f-a4d4-c92a1999eaad-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485778 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pr2h\" (UniqueName: \"kubernetes.io/projected/0b337efa-c799-4fce-833c-cd0149c3644c-kube-api-access-9pr2h\") pod \"migrator-59844c95c7-x8sls\" (UID: \"0b337efa-c799-4fce-833c-cd0149c3644c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485864 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm5lq\" (UniqueName: \"kubernetes.io/projected/24abe58b-4b74-4207-ba8f-1b16c22617cc-kube-api-access-bm5lq\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485889 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-plugins-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485912 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8v4q\" (UniqueName: \"kubernetes.io/projected/5b11a69d-9e3a-4e77-8c17-a9f2cc34873d-kube-api-access-k8v4q\") pod \"ingress-canary-29qfq\" (UID: \"5b11a69d-9e3a-4e77-8c17-a9f2cc34873d\") " pod="openshift-ingress-canary/ingress-canary-29qfq" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485981 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.485997 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjr97\" (UniqueName: \"kubernetes.io/projected/ee4b8c43-5789-4d06-884e-f2fcc355eeec-kube-api-access-hjr97\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486042 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-bound-sa-token\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486068 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d143445-9a26-4203-995d-82238aa12f54-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4f9mb\" (UID: \"9d143445-9a26-4203-995d-82238aa12f54\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486084 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59clc\" (UniqueName: \"kubernetes.io/projected/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-kube-api-access-59clc\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486104 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gftn\" (UniqueName: \"kubernetes.io/projected/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-kube-api-access-8gftn\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486130 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78372347-a30d-4a0f-a4d4-c92a1999eaad-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486150 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24abe58b-4b74-4207-ba8f-1b16c22617cc-config-volume\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486165 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nln87\" (UniqueName: \"kubernetes.io/projected/b209aa10-3135-455f-8ad0-a54f64b7768d-kube-api-access-nln87\") pod \"multus-admission-controller-857f4d67dd-frs8b\" (UID: \"b209aa10-3135-455f-8ad0-a54f64b7768d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486181 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vv9n\" (UniqueName: \"kubernetes.io/projected/0784a997-5d78-4199-a3fa-3360eef7a756-kube-api-access-2vv9n\") pod \"package-server-manager-789f6589d5-8c9n2\" (UID: \"0784a997-5d78-4199-a3fa-3360eef7a756\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486213 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24abe58b-4b74-4207-ba8f-1b16c22617cc-secret-volume\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486229 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-signing-key\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486255 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486295 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-csi-data-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486311 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-config-volume\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486370 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-tls\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486403 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b11a69d-9e3a-4e77-8c17-a9f2cc34873d-cert\") pod \"ingress-canary-29qfq\" (UID: \"5b11a69d-9e3a-4e77-8c17-a9f2cc34873d\") " pod="openshift-ingress-canary/ingress-canary-29qfq" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486437 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-certificates\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486476 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmjpk\" (UniqueName: \"kubernetes.io/projected/617d3d62-5b4e-481b-817c-54288406fc7c-kube-api-access-dmjpk\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486814 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-trusted-ca\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.486974 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glx8g\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-kube-api-access-glx8g\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.488264 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-csi-data-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.489313 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-config-volume\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.493315 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78372347-a30d-4a0f-a4d4-c92a1999eaad-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.494210 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-plugins-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.495525 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-mountpoint-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.498758 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-trusted-ca\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.499213 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24abe58b-4b74-4207-ba8f-1b16c22617cc-config-volume\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.499301 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:21.999279261 +0000 UTC m=+155.846954909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.499476 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-certificates\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.500676 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-registration-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.501056 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.501134 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/617d3d62-5b4e-481b-817c-54288406fc7c-socket-dir\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.501920 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-signing-cabundle\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.502724 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.503562 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b209aa10-3135-455f-8ad0-a54f64b7768d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-frs8b\" (UID: \"b209aa10-3135-455f-8ad0-a54f64b7768d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.508910 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/26b1f854-1ae7-4987-9454-94263ed6b5d7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.510047 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d143445-9a26-4203-995d-82238aa12f54-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4f9mb\" (UID: \"9d143445-9a26-4203-995d-82238aa12f54\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.510632 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78372347-a30d-4a0f-a4d4-c92a1999eaad-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.510840 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-tls\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.515110 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0784a997-5d78-4199-a3fa-3360eef7a756-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8c9n2\" (UID: \"0784a997-5d78-4199-a3fa-3360eef7a756\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.534053 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-metrics-tls\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.534444 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/26b1f854-1ae7-4987-9454-94263ed6b5d7-srv-cert\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.535897 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24abe58b-4b74-4207-ba8f-1b16c22617cc-secret-volume\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.536141 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-signing-key\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.536880 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b11a69d-9e3a-4e77-8c17-a9f2cc34873d-cert\") pod \"ingress-canary-29qfq\" (UID: \"5b11a69d-9e3a-4e77-8c17-a9f2cc34873d\") " pod="openshift-ingress-canary/ingress-canary-29qfq" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.552250 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glx8g\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-kube-api-access-glx8g\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.552304 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr"] Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.553562 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fpq7g"] Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.558176 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-bound-sa-token\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.576222 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vv9n\" (UniqueName: \"kubernetes.io/projected/0784a997-5d78-4199-a3fa-3360eef7a756-kube-api-access-2vv9n\") pod \"package-server-manager-789f6589d5-8c9n2\" (UID: \"0784a997-5d78-4199-a3fa-3360eef7a756\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.587975 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.588636 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.088587153 +0000 UTC m=+155.936262621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.593424 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59clc\" (UniqueName: \"kubernetes.io/projected/7206a358-c3d3-4f1b-ac51-4c0657f9e29a-kube-api-access-59clc\") pod \"dns-default-blzx5\" (UID: \"7206a358-c3d3-4f1b-ac51-4c0657f9e29a\") " pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.648212 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gftn\" (UniqueName: \"kubernetes.io/projected/c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3-kube-api-access-8gftn\") pod \"service-ca-9c57cc56f-78r68\" (UID: \"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3\") " pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.658955 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nln87\" (UniqueName: \"kubernetes.io/projected/b209aa10-3135-455f-8ad0-a54f64b7768d-kube-api-access-nln87\") pod \"multus-admission-controller-857f4d67dd-frs8b\" (UID: \"b209aa10-3135-455f-8ad0-a54f64b7768d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.672748 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm5lq\" (UniqueName: \"kubernetes.io/projected/24abe58b-4b74-4207-ba8f-1b16c22617cc-kube-api-access-bm5lq\") pod \"collect-profiles-29325870-cvw6n\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.684527 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8v4q\" (UniqueName: \"kubernetes.io/projected/5b11a69d-9e3a-4e77-8c17-a9f2cc34873d-kube-api-access-k8v4q\") pod \"ingress-canary-29qfq\" (UID: \"5b11a69d-9e3a-4e77-8c17-a9f2cc34873d\") " pod="openshift-ingress-canary/ingress-canary-29qfq" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.690326 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.690858 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.19084569 +0000 UTC m=+156.038521148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.692945 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-78r68" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.697043 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.700216 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.704020 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmjpk\" (UniqueName: \"kubernetes.io/projected/617d3d62-5b4e-481b-817c-54288406fc7c-kube-api-access-dmjpk\") pod \"csi-hostpathplugin-ttqhz\" (UID: \"617d3d62-5b4e-481b-817c-54288406fc7c\") " pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.720847 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc24k\" (UniqueName: \"kubernetes.io/projected/26b1f854-1ae7-4987-9454-94263ed6b5d7-kube-api-access-fc24k\") pod \"olm-operator-6b444d44fb-z2mg2\" (UID: \"26b1f854-1ae7-4987-9454-94263ed6b5d7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.721512 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.755230 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj"] Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.763651 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pr2h\" (UniqueName: \"kubernetes.io/projected/0b337efa-c799-4fce-833c-cd0149c3644c-kube-api-access-9pr2h\") pod \"migrator-59844c95c7-x8sls\" (UID: \"0b337efa-c799-4fce-833c-cd0149c3644c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.766804 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fssv5\" (UniqueName: \"kubernetes.io/projected/9d143445-9a26-4203-995d-82238aa12f54-kube-api-access-fssv5\") pod \"control-plane-machine-set-operator-78cbb6b69f-4f9mb\" (UID: \"9d143445-9a26-4203-995d-82238aa12f54\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.775879 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf"] Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.775951 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjr97\" (UniqueName: \"kubernetes.io/projected/ee4b8c43-5789-4d06-884e-f2fcc355eeec-kube-api-access-hjr97\") pod \"marketplace-operator-79b997595-tz4gp\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.791212 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.791575 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.291561046 +0000 UTC m=+156.139236514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:21 crc kubenswrapper[4992]: W1004 04:35:21.835571 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b5baf7a_852f_47b1_be0b_3daf981d2281.slice/crio-e10fd1bfb228ef77a39e2b402cd85659ed7e89fd6f32e3dcd91f22d60e7f4161 WatchSource:0}: Error finding container e10fd1bfb228ef77a39e2b402cd85659ed7e89fd6f32e3dcd91f22d60e7f4161: Status 404 returned error can't find the container with id e10fd1bfb228ef77a39e2b402cd85659ed7e89fd6f32e3dcd91f22d60e7f4161 Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.891890 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.894620 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.895073 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.395057386 +0000 UTC m=+156.242732844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.907569 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.917244 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.949470 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.959837 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-29qfq" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.974917 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.990723 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" Oct 04 04:35:21 crc kubenswrapper[4992]: I1004 04:35:21.995968 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:21 crc kubenswrapper[4992]: E1004 04:35:21.996896 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.496875981 +0000 UTC m=+156.344551449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.044257 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" event={"ID":"fc9173a7-e308-408a-896b-e2990f50d8c6","Type":"ContainerStarted","Data":"6e0e8b65df910a1bcca5d5b867f9b065d55a48b05897dbf32d6c98bb174bd53d"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.049530 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" event={"ID":"3d028a5e-d0df-44c7-8e0a-73949ccb3984","Type":"ContainerStarted","Data":"32b0a3ca4953ed4ac886aa958a7e1b7b7d02e223713d3badeba9a939a0773ab9"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.061511 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" event={"ID":"2fa6af32-186a-42ac-9513-398cd1124410","Type":"ContainerStarted","Data":"709068e86247c0f449c8cfb79272505b9cbd5ae234103c221b4072021e44cd90"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.061565 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" event={"ID":"2fa6af32-186a-42ac-9513-398cd1124410","Type":"ContainerStarted","Data":"80a93eed1d902d287072af68f80d072b57cca2652bfd51682f7be3bd389efa66"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.064618 4992 generic.go:334] "Generic (PLEG): container finished" podID="3cd8a6e8-ee96-4e7e-9367-e74478678f5f" containerID="85695c8c6625a141ba9fbaf82a35a5d071321f8987739c1dd4ea3b6130728db0" exitCode=0 Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.064667 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" event={"ID":"3cd8a6e8-ee96-4e7e-9367-e74478678f5f","Type":"ContainerDied","Data":"85695c8c6625a141ba9fbaf82a35a5d071321f8987739c1dd4ea3b6130728db0"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.068916 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" event={"ID":"1b5baf7a-852f-47b1-be0b-3daf981d2281","Type":"ContainerStarted","Data":"e10fd1bfb228ef77a39e2b402cd85659ed7e89fd6f32e3dcd91f22d60e7f4161"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.086346 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" event={"ID":"b240811c-d6e5-4a4f-932a-fdf41ad86a98","Type":"ContainerStarted","Data":"fea666f67cf0fd6ef9abb82415c79f333d5aa7f94c0830f64b260765ba413a50"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.100831 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" event={"ID":"06f9131b-3fbf-468f-8201-13cc313f7f6c","Type":"ContainerStarted","Data":"ee5e98c462dd1e9e05ae92add4529bea997f3af29a7371be0042fbc8e7939af0"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.101257 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.101779 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.601764128 +0000 UTC m=+156.449439596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.102208 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" event={"ID":"5f3bb363-9319-469e-aca0-29eb769f92ec","Type":"ContainerStarted","Data":"322da91d9280c1f91ffe677138c1e6f5b21dc8a8b69a1f1e0446b6111bc5e077"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.106621 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" event={"ID":"e1a86390-9e3f-4ddf-8440-40b2e4d88412","Type":"ContainerStarted","Data":"2e39434b63291655372749061605c2700a6bf41ace8f615733272d52c6e9d29d"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.106656 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" event={"ID":"e1a86390-9e3f-4ddf-8440-40b2e4d88412","Type":"ContainerStarted","Data":"f978846948251a95c6af98d90a4b2319177e606f30ef536412592c38eb778688"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.143208 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" event={"ID":"c2c1be78-1a57-47ca-a671-cb0a43e05c2a","Type":"ContainerStarted","Data":"d7299ea91f203d27f8d62907045f69af3d2cff292198a0a9e43b075e09048159"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.153042 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5rtlc" event={"ID":"62b435a3-f5a3-4cef-a80c-5968f148d816","Type":"ContainerStarted","Data":"12d2f8f255d952d7121a680966da151915304cfb367c2a086e99bfa0bfc51151"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.159713 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-68nfv" event={"ID":"80d7820e-ea64-433e-9842-9142a578f408","Type":"ContainerStarted","Data":"fe2d7fc10307be5766d96835a748690320a91a8dd9c7b35b062e34ef61ccdac9"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.163948 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" event={"ID":"d8c6ed95-6bcb-4a2d-b7d4-de1419159770","Type":"ContainerStarted","Data":"a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df"} Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.164828 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.173431 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.178518 4992 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-fb7m9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.178586 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" podUID="d8c6ed95-6bcb-4a2d-b7d4-de1419159770" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.189204 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hmx5v"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.202499 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.208432 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.708399951 +0000 UTC m=+156.556075589 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.313715 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.314172 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.315375 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.815340702 +0000 UTC m=+156.663016170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: W1004 04:35:22.316695 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55b54660_afa7_4a76_b86c_9d1fd3f33769.slice/crio-fa6dd2f4c0d07e36a83f70a31e3ea84b075755b0d311573399a44f0d522d4d98 WatchSource:0}: Error finding container fa6dd2f4c0d07e36a83f70a31e3ea84b075755b0d311573399a44f0d522d4d98: Status 404 returned error can't find the container with id fa6dd2f4c0d07e36a83f70a31e3ea84b075755b0d311573399a44f0d522d4d98 Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.318084 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config\") pod \"route-controller-manager-6576b87f9c-4tqf8\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.415266 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.416439 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:22.916419688 +0000 UTC m=+156.764095166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.470658 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.517149 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.518196 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.018178802 +0000 UTC m=+156.865854270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.629072 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.629581 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.129563291 +0000 UTC m=+156.977238759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.693803 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rd62r"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.706693 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-79tvb"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.710741 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.717107 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5sgbq"] Oct 04 04:35:22 crc kubenswrapper[4992]: W1004 04:35:22.718348 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd501849_1d14_48c1_bc51_06d23c56db28.slice/crio-895c15af4ca9601edb7369b96e0076e104933d1e855eab74dbb7799280877b72 WatchSource:0}: Error finding container 895c15af4ca9601edb7369b96e0076e104933d1e855eab74dbb7799280877b72: Status 404 returned error can't find the container with id 895c15af4ca9601edb7369b96e0076e104933d1e855eab74dbb7799280877b72 Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.719972 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r7t8c"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.722761 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pjtjn"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.731855 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.732860 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.232844985 +0000 UTC m=+157.080520453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: W1004 04:35:22.741105 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod533ee8bc_a31e_43a7_b87c_911ecd5c191d.slice/crio-83da047203422579e1c1eb95fa0eeaeefe65f60e4a260eb75411a91b213a9893 WatchSource:0}: Error finding container 83da047203422579e1c1eb95fa0eeaeefe65f60e4a260eb75411a91b213a9893: Status 404 returned error can't find the container with id 83da047203422579e1c1eb95fa0eeaeefe65f60e4a260eb75411a91b213a9893 Oct 04 04:35:22 crc kubenswrapper[4992]: W1004 04:35:22.748107 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e339a50_fbf4_44da_9335_ba6ae033fc38.slice/crio-b24cf50aaf443a609f4671a6ceddb2b2af964f93c201e12ec69840dbe47ad7aa WatchSource:0}: Error finding container b24cf50aaf443a609f4671a6ceddb2b2af964f93c201e12ec69840dbe47ad7aa: Status 404 returned error can't find the container with id b24cf50aaf443a609f4671a6ceddb2b2af964f93c201e12ec69840dbe47ad7aa Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.833104 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.833796 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.333776147 +0000 UTC m=+157.181451625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: W1004 04:35:22.864785 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod103e81ce_51a3_495a_adf7_87e8af781036.slice/crio-a72147881813ce9c9481f057cb92a92bed3440f4d1645b34eb694372f5a1a3ba WatchSource:0}: Error finding container a72147881813ce9c9481f057cb92a92bed3440f4d1645b34eb694372f5a1a3ba: Status 404 returned error can't find the container with id a72147881813ce9c9481f057cb92a92bed3440f4d1645b34eb694372f5a1a3ba Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.890938 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.897481 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.903970 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.912831 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.932928 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.934872 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:22 crc kubenswrapper[4992]: E1004 04:35:22.935222 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.435207752 +0000 UTC m=+157.282883220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.952429 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.953164 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-blzx5"] Oct 04 04:35:22 crc kubenswrapper[4992]: I1004 04:35:22.955245 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw"] Oct 04 04:35:22 crc kubenswrapper[4992]: W1004 04:35:22.988513 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78b0d5e3_fbae_4a2b_9ddf_6c8cdbdb319d.slice/crio-8be88df559ec27d8f9d4beabcb9d3837fb229bf3d51cafde972773731096c81c WatchSource:0}: Error finding container 8be88df559ec27d8f9d4beabcb9d3837fb229bf3d51cafde972773731096c81c: Status 404 returned error can't find the container with id 8be88df559ec27d8f9d4beabcb9d3837fb229bf3d51cafde972773731096c81c Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.038981 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.045635 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.545593255 +0000 UTC m=+157.393268733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: W1004 04:35:23.075603 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaee1200b_878a_4a90_a34c_dcc237c6a8a3.slice/crio-5e89bd6dbcf55f89bdb6ecef359d8007fbb34eca398ca05a79a92fdcb0c8139e WatchSource:0}: Error finding container 5e89bd6dbcf55f89bdb6ecef359d8007fbb34eca398ca05a79a92fdcb0c8139e: Status 404 returned error can't find the container with id 5e89bd6dbcf55f89bdb6ecef359d8007fbb34eca398ca05a79a92fdcb0c8139e Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.102586 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.118529 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k5knb"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.140403 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.140864 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.640850165 +0000 UTC m=+157.488525633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.158675 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.173639 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.192400 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-78r68"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.192519 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.202682 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" event={"ID":"9e540236-2888-4509-a53e-9e51cae33546","Type":"ContainerStarted","Data":"36ca27bd32e64a4d206178cdf7225b944fda629af5f23436610cb94b5e7998c4"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.213531 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.213763 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" event={"ID":"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d","Type":"ContainerStarted","Data":"8be88df559ec27d8f9d4beabcb9d3837fb229bf3d51cafde972773731096c81c"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.221883 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-29qfq"] Oct 04 04:35:23 crc kubenswrapper[4992]: W1004 04:35:23.222194 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35350b4a_c3a6_4d70_bab2_a79757701029.slice/crio-25d9378f456f67f42d25644775466d6dee0c156a7b5b72358e83fd76b06a97ca WatchSource:0}: Error finding container 25d9378f456f67f42d25644775466d6dee0c156a7b5b72358e83fd76b06a97ca: Status 404 returned error can't find the container with id 25d9378f456f67f42d25644775466d6dee0c156a7b5b72358e83fd76b06a97ca Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.225545 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" podStartSLOduration=127.225523015 podStartE2EDuration="2m7.225523015s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.21141722 +0000 UTC m=+157.059092698" watchObservedRunningTime="2025-10-04 04:35:23.225523015 +0000 UTC m=+157.073198483" Oct 04 04:35:23 crc kubenswrapper[4992]: W1004 04:35:23.226828 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0784a997_5d78_4199_a3fa_3360eef7a756.slice/crio-b7d2e16d4997bd3685dfe7a28f33e5c1185cd06617c55275ea78558c2fd4c281 WatchSource:0}: Error finding container b7d2e16d4997bd3685dfe7a28f33e5c1185cd06617c55275ea78558c2fd4c281: Status 404 returned error can't find the container with id b7d2e16d4997bd3685dfe7a28f33e5c1185cd06617c55275ea78558c2fd4c281 Oct 04 04:35:23 crc kubenswrapper[4992]: W1004 04:35:23.227121 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24abe58b_4b74_4207_ba8f_1b16c22617cc.slice/crio-bbb3f2f79ddb48c6a537f07202b8b68e91152e82a86798e33b37ed7b51e43091 WatchSource:0}: Error finding container bbb3f2f79ddb48c6a537f07202b8b68e91152e82a86798e33b37ed7b51e43091: Status 404 returned error can't find the container with id bbb3f2f79ddb48c6a537f07202b8b68e91152e82a86798e33b37ed7b51e43091 Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.227551 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" event={"ID":"dd501849-1d14-48c1-bc51-06d23c56db28","Type":"ContainerStarted","Data":"895c15af4ca9601edb7369b96e0076e104933d1e855eab74dbb7799280877b72"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.232120 4992 generic.go:334] "Generic (PLEG): container finished" podID="82d8b4e3-0fe2-4941-895e-47513a81b9bf" containerID="b7cd49a269744d66a9a06c77cde9cc9a4be4be9a1c3b68c94f319dcf705bfbea" exitCode=0 Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.232206 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" event={"ID":"82d8b4e3-0fe2-4941-895e-47513a81b9bf","Type":"ContainerDied","Data":"b7cd49a269744d66a9a06c77cde9cc9a4be4be9a1c3b68c94f319dcf705bfbea"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.232257 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" event={"ID":"82d8b4e3-0fe2-4941-895e-47513a81b9bf","Type":"ContainerStarted","Data":"b976dfbc92743f29cfa8222b42f2a197e7be05d1ae7263c72e97dbae59521063"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.237080 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" event={"ID":"55b54660-afa7-4a76-b86c-9d1fd3f33769","Type":"ContainerStarted","Data":"c4ee38a2b625aed52fd9a8d71d64f2b8aa028ce2313fd94add8924a4ed85fe01"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.237145 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" event={"ID":"55b54660-afa7-4a76-b86c-9d1fd3f33769","Type":"ContainerStarted","Data":"fa6dd2f4c0d07e36a83f70a31e3ea84b075755b0d311573399a44f0d522d4d98"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.238119 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.241212 4992 patch_prober.go:28] interesting pod/console-operator-58897d9998-hmx5v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.241294 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" podUID="55b54660-afa7-4a76-b86c-9d1fd3f33769" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.241674 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.242235 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.742219709 +0000 UTC m=+157.589895177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.247134 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.248169 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r7t8c" event={"ID":"6a9ef22d-4511-403a-9703-fcffd8f5b8e9","Type":"ContainerStarted","Data":"8870199d64e3d2a0ae1cee00dd206b8ddd3ccd7ca374dec2962bc0c6bab9e05a"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.266024 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5rtlc" event={"ID":"62b435a3-f5a3-4cef-a80c-5968f148d816","Type":"ContainerStarted","Data":"f0756f51d74b7c65cb8a34451d31f18141978b0bded20fc494081432c93e2265"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.269177 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-68nfv" event={"ID":"80d7820e-ea64-433e-9842-9142a578f408","Type":"ContainerStarted","Data":"ce6370f4aded2f2ab816e1a05b77b2352adf719a16f2302d3cb9a0c24bd173f9"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.273906 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-blzx5" event={"ID":"7206a358-c3d3-4f1b-ac51-4c0657f9e29a","Type":"ContainerStarted","Data":"7b486406578f38ede7cbf2a84c86ae2147a4b64734fdb3c89331c83cea5b85bb"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.275706 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" event={"ID":"aee1200b-878a-4a90-a34c-dcc237c6a8a3","Type":"ContainerStarted","Data":"5e89bd6dbcf55f89bdb6ecef359d8007fbb34eca398ca05a79a92fdcb0c8139e"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.276998 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" event={"ID":"f796a803-cdb3-46bf-82f0-9e89f7f50a4a","Type":"ContainerStarted","Data":"b8fe0483911d019665093b1fe2c4c9a5112b9344b7c82e9d264f16003e88c65a"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.280618 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" event={"ID":"86a9fb33-b1f8-4f50-9785-a5504e4b4efd","Type":"ContainerStarted","Data":"95422341d59b09930d23aab2f991fe6b4699bed079179b3d75496b0797bd07f1"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.286229 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.288256 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" event={"ID":"1b5baf7a-852f-47b1-be0b-3daf981d2281","Type":"ContainerStarted","Data":"60bd3d3b6cdbe8eccc58b5b464f210cb1c27c5f33b0dcec97fd766049f6d6b66"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.288309 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" event={"ID":"1b5baf7a-852f-47b1-be0b-3daf981d2281","Type":"ContainerStarted","Data":"8765f1e91f7f2717628b2339bc8945c7a8b0a7b70c108037afc44d132934d393"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.295024 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" event={"ID":"b240811c-d6e5-4a4f-932a-fdf41ad86a98","Type":"ContainerStarted","Data":"0610efea3c941c58eecc4611aaaf4166ca40c30676951c1ecacb5179c4679316"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.295621 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.295608 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tbcnm" podStartSLOduration=128.295587167 podStartE2EDuration="2m8.295587167s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.249623225 +0000 UTC m=+157.097298713" watchObservedRunningTime="2025-10-04 04:35:23.295587167 +0000 UTC m=+157.143262635" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.295660 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.296443 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-cxgpw" podStartSLOduration=128.296436429 podStartE2EDuration="2m8.296436429s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.294459467 +0000 UTC m=+157.142134955" watchObservedRunningTime="2025-10-04 04:35:23.296436429 +0000 UTC m=+157.144111897" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.306921 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" event={"ID":"20f97015-5d3a-43a3-8863-95fcc8dd8414","Type":"ContainerStarted","Data":"d377786ace77e69e6829a68f6102417f31714e1f6267db0e02d9e72c6a65e035"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.308960 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tz4gp"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.311006 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" event={"ID":"fc9173a7-e308-408a-896b-e2990f50d8c6","Type":"ContainerStarted","Data":"63b516bb74f60e86931b3e4e48324a509c1eaa9ae67449b839c3d97a8e35bffb"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.311044 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" event={"ID":"fc9173a7-e308-408a-896b-e2990f50d8c6","Type":"ContainerStarted","Data":"298af8b191973e99de4e8d5d2d9d97fd9487e8705993ffc4132da49512e47d28"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.324799 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-frs8b"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.328005 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" event={"ID":"533ee8bc-a31e-43a7-b87c-911ecd5c191d","Type":"ContainerStarted","Data":"83da047203422579e1c1eb95fa0eeaeefe65f60e4a260eb75411a91b213a9893"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.336525 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" event={"ID":"5f3bb363-9319-469e-aca0-29eb769f92ec","Type":"ContainerStarted","Data":"95f791dbc209c66c0cd0ca0fec675c935a746b33b0f93490accfac1a154e9552"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.345123 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.345163 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.345206 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.345234 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.345309 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.345576 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5sgbq" event={"ID":"8e339a50-fbf4-44da-9335-ba6ae033fc38","Type":"ContainerStarted","Data":"b24cf50aaf443a609f4671a6ceddb2b2af964f93c201e12ec69840dbe47ad7aa"} Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.347093 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.847077675 +0000 UTC m=+157.694753353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.350712 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" event={"ID":"60d29eac-b443-4754-b9bb-d21591d2333d","Type":"ContainerStarted","Data":"2552462b4bf6824b59a52e00e32682eb3d0bc6f29e9e959e41fcc713e11e81c6"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.353798 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.355838 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" podStartSLOduration=128.355823707 podStartE2EDuration="2m8.355823707s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.348573704 +0000 UTC m=+157.196249172" watchObservedRunningTime="2025-10-04 04:35:23.355823707 +0000 UTC m=+157.203499175" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.359915 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ttqhz"] Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.359961 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" event={"ID":"3cd8a6e8-ee96-4e7e-9367-e74478678f5f","Type":"ContainerStarted","Data":"62b3bbd2b2c659e44b4b9e2c80f1edf02c43efffa633bca3d5427311754fd227"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.362138 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.370391 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.372007 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" event={"ID":"103e81ce-51a3-495a-adf7-87e8af781036","Type":"ContainerStarted","Data":"a72147881813ce9c9481f057cb92a92bed3440f4d1645b34eb694372f5a1a3ba"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.373237 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.376935 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" event={"ID":"c2c1be78-1a57-47ca-a671-cb0a43e05c2a","Type":"ContainerStarted","Data":"679152e9cce457fba862d8c78860cb6d49d25c8c8401846415ed89621f924f41"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.384581 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" event={"ID":"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8","Type":"ContainerStarted","Data":"85a72c8ac3f883b90987aa920ac2e5e3b162abbc92053e43f1f3c3b63a5a4d2e"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.389204 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" event={"ID":"06f9131b-3fbf-468f-8201-13cc313f7f6c","Type":"ContainerStarted","Data":"fdb739ea4ab5334f0249defa77cc8610997b827eed191a31a4cb750a1e9882b0"} Oct 04 04:35:23 crc kubenswrapper[4992]: W1004 04:35:23.390577 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee4b8c43_5789_4d06_884e_f2fcc355eeec.slice/crio-f31a06270abf47b7476852f7b72a87ef19a2897f394eba9ac188538f806bbf14 WatchSource:0}: Error finding container f31a06270abf47b7476852f7b72a87ef19a2897f394eba9ac188538f806bbf14: Status 404 returned error can't find the container with id f31a06270abf47b7476852f7b72a87ef19a2897f394eba9ac188538f806bbf14 Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.390995 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" event={"ID":"3d028a5e-d0df-44c7-8e0a-73949ccb3984","Type":"ContainerStarted","Data":"412d923d9e2d5b3224b10e9155d8e03318aa7222a710c3fa121fabc75d8eb0a4"} Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.392466 4992 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-fb7m9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.392504 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" podUID="d8c6ed95-6bcb-4a2d-b7d4-de1419159770" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.415648 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-68nfv" podStartSLOduration=127.415626076 podStartE2EDuration="2m7.415626076s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.415173514 +0000 UTC m=+157.262848992" watchObservedRunningTime="2025-10-04 04:35:23.415626076 +0000 UTC m=+157.263301544" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.446184 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.446495 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.946467655 +0000 UTC m=+157.794143253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.446727 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.448775 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:23.948759826 +0000 UTC m=+157.796435494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.497897 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5rtlc" podStartSLOduration=5.497865271 podStartE2EDuration="5.497865271s" podCreationTimestamp="2025-10-04 04:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.447487432 +0000 UTC m=+157.295162900" watchObservedRunningTime="2025-10-04 04:35:23.497865271 +0000 UTC m=+157.345540739" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.498385 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9kblj" podStartSLOduration=127.498379795 podStartE2EDuration="2m7.498379795s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.496215247 +0000 UTC m=+157.343890715" watchObservedRunningTime="2025-10-04 04:35:23.498379795 +0000 UTC m=+157.346055263" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.530375 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h86m4" podStartSLOduration=129.530337154 podStartE2EDuration="2m9.530337154s" podCreationTimestamp="2025-10-04 04:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.53018979 +0000 UTC m=+157.377865258" watchObservedRunningTime="2025-10-04 04:35:23.530337154 +0000 UTC m=+157.378012622" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.544676 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.549069 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.549632 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.049602936 +0000 UTC m=+157.897278404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.549768 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.552506 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.052482062 +0000 UTC m=+157.900157730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.609084 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" podStartSLOduration=127.609060535 podStartE2EDuration="2m7.609060535s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.571393855 +0000 UTC m=+157.419069333" watchObservedRunningTime="2025-10-04 04:35:23.609060535 +0000 UTC m=+157.456736003" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.634127 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.644783 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.651374 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.151325458 +0000 UTC m=+157.999000926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.651963 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-w4dg6" podStartSLOduration=128.651943485 podStartE2EDuration="2m8.651943485s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.610219576 +0000 UTC m=+157.457895044" watchObservedRunningTime="2025-10-04 04:35:23.651943485 +0000 UTC m=+157.499618963" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.652087 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.652210 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8zxrr" podStartSLOduration=127.652204612 podStartE2EDuration="2m7.652204612s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.650539647 +0000 UTC m=+157.498215135" watchObservedRunningTime="2025-10-04 04:35:23.652204612 +0000 UTC m=+157.499880080" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.652475 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.652901 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.15288511 +0000 UTC m=+158.000560578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.701084 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gh7xf" podStartSLOduration=127.701015258 podStartE2EDuration="2m7.701015258s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.691345551 +0000 UTC m=+157.539021019" watchObservedRunningTime="2025-10-04 04:35:23.701015258 +0000 UTC m=+157.548690716" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.735419 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-fpq7g" podStartSLOduration=127.735385201 podStartE2EDuration="2m7.735385201s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.734948099 +0000 UTC m=+157.582623577" watchObservedRunningTime="2025-10-04 04:35:23.735385201 +0000 UTC m=+157.583060689" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.753938 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.754404 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.254382755 +0000 UTC m=+158.102058223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.775047 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-pbx8v" podStartSLOduration=127.775013184 podStartE2EDuration="2m7.775013184s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:23.770615057 +0000 UTC m=+157.618290525" watchObservedRunningTime="2025-10-04 04:35:23.775013184 +0000 UTC m=+157.622688652" Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.863461 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.863996 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.363973307 +0000 UTC m=+158.211648945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:23 crc kubenswrapper[4992]: I1004 04:35:23.964536 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:23 crc kubenswrapper[4992]: E1004 04:35:23.965597 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.465573416 +0000 UTC m=+158.313248884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.067024 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.067689 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.567670489 +0000 UTC m=+158.415345957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.169065 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.170154 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.670136042 +0000 UTC m=+158.517811510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.277231 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.277634 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.777617367 +0000 UTC m=+158.625292835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.291119 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:24 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:24 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:24 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.291167 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.379505 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.380201 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.880181552 +0000 UTC m=+158.727857020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.443053 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" event={"ID":"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8","Type":"ContainerStarted","Data":"9232aab7bae4d108455b0fceecee999983fd9ca1d3e76209ad972b60f6ca31e0"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.448045 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" event={"ID":"b209aa10-3135-455f-8ad0-a54f64b7768d","Type":"ContainerStarted","Data":"f4e4f182f6c3b05a96f1d89d6d40641e6af29d6c3df1756b2750aab5f45c101f"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.481433 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5sgbq" event={"ID":"8e339a50-fbf4-44da-9335-ba6ae033fc38","Type":"ContainerStarted","Data":"19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.488228 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.488710 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:24.988693215 +0000 UTC m=+158.836368683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.500043 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" event={"ID":"60d29eac-b443-4754-b9bb-d21591d2333d","Type":"ContainerStarted","Data":"ca0ff5620ab143d8ce53181802e4a780d592dbc2f9691530a5b2ebcd134b9301"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.591142 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.610329 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.110299666 +0000 UTC m=+158.957975134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.611043 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.617344 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.117318263 +0000 UTC m=+158.964993731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.640593 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" event={"ID":"dd501849-1d14-48c1-bc51-06d23c56db28","Type":"ContainerStarted","Data":"c0e948ba0c4d699e3726f94988295b17571d1279775fbe46aa70fe01836a16f7"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.688686 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.692941 4992 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-rd62r container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.693005 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" podUID="dd501849-1d14-48c1-bc51-06d23c56db28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.704159 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" event={"ID":"26b1f854-1ae7-4987-9454-94263ed6b5d7","Type":"ContainerStarted","Data":"ca1b0e449a2ec1687d40e3580d7be9922382010e311d588a2800be0d976f98a4"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.704210 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" event={"ID":"26b1f854-1ae7-4987-9454-94263ed6b5d7","Type":"ContainerStarted","Data":"f916b259093dfb9785e2f88fe1ee4d482d5ace7589ad12cfbaa795ec7595d8ce"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.704633 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.712209 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.713308 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.213291903 +0000 UTC m=+159.060967371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.714220 4992 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-z2mg2 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.714297 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" podUID="26b1f854-1ae7-4987-9454-94263ed6b5d7" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.743168 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" event={"ID":"35350b4a-c3a6-4d70-bab2-a79757701029","Type":"ContainerStarted","Data":"25d9378f456f67f42d25644775466d6dee0c156a7b5b72358e83fd76b06a97ca"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.760506 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" event={"ID":"9e540236-2888-4509-a53e-9e51cae33546","Type":"ContainerStarted","Data":"fcd19e5539dcf0eb8dbf92465ee5435c1adfbc1afb2beeb7233abe337c273a5b"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.830316 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.832110 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.332093889 +0000 UTC m=+159.179769357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.867911 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" event={"ID":"aee1200b-878a-4a90-a34c-dcc237c6a8a3","Type":"ContainerStarted","Data":"b1a9b942bf81ec4c8a5f8903fc3209779e0176f11603f660af1dea37387b59b0"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.869230 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.884226 4992 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-xjmbw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.884286 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" podUID="aee1200b-878a-4a90-a34c-dcc237c6a8a3" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.912309 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" event={"ID":"f796a803-cdb3-46bf-82f0-9e89f7f50a4a","Type":"ContainerStarted","Data":"ff31d2b6e0f69d9c713da3f0694b9446ab15fe831e0dcf1ebe1eb9867169b5aa"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.935917 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" event={"ID":"4256fad3-f7d5-4441-a4f0-ee84712d7321","Type":"ContainerStarted","Data":"2d17d269865c95134c47ae932655df0ca28b9147f303af4098d5bb20d70cab59"} Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.937928 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.941705 4992 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4tqf8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.941772 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" podUID="4256fad3-f7d5-4441-a4f0-ee84712d7321" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.944814 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.944952 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.444928287 +0000 UTC m=+159.292603755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:24 crc kubenswrapper[4992]: I1004 04:35:24.945141 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:24 crc kubenswrapper[4992]: E1004 04:35:24.946153 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.446141939 +0000 UTC m=+159.293817407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.000303 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" event={"ID":"617d3d62-5b4e-481b-817c-54288406fc7c","Type":"ContainerStarted","Data":"ffa4a94d5bbccee47e6574cf97173d155d963ca47e4610909a90d1d76c5ae734"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.014631 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r7t8c" event={"ID":"6a9ef22d-4511-403a-9703-fcffd8f5b8e9","Type":"ContainerStarted","Data":"1371c77d74650e0426a7b16f479a7a9595aa8c05934cc60da8142063f6b18f32"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.015452 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.019900 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.019962 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.047172 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.048771 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.548755865 +0000 UTC m=+159.396431333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.060342 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" event={"ID":"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d","Type":"ContainerStarted","Data":"a846beb2cab270cb09c7ad0dc04a66ac60d7a052a4e2117b4d9f95339d590d9c"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.081870 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" event={"ID":"0784a997-5d78-4199-a3fa-3360eef7a756","Type":"ContainerStarted","Data":"a235316e17a09c19d04502dc5b301c2b1915973df6de04f6d6b4a4d9dcb3654a"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.081939 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" event={"ID":"0784a997-5d78-4199-a3fa-3360eef7a756","Type":"ContainerStarted","Data":"b7d2e16d4997bd3685dfe7a28f33e5c1185cd06617c55275ea78558c2fd4c281"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.113827 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" event={"ID":"82d8b4e3-0fe2-4941-895e-47513a81b9bf","Type":"ContainerStarted","Data":"375cf0d3037babdf536f8d04164255db32708bf7158b424f81754b8b2c66c557"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.114468 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.150934 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.151311 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.65129582 +0000 UTC m=+159.498971288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.161111 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" event={"ID":"20f97015-5d3a-43a3-8863-95fcc8dd8414","Type":"ContainerStarted","Data":"1dde9d77028a3849681cd37d04ee3dc68c645f4d3047955c4087cbeef48931c6"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.168637 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-blzx5" event={"ID":"7206a358-c3d3-4f1b-ac51-4c0657f9e29a","Type":"ContainerStarted","Data":"5a20a2fb0d76f079c7bdd6fe9cfc43ab782b7d9504d66a44984b4ece7edc4f3b"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.196069 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" event={"ID":"9d143445-9a26-4203-995d-82238aa12f54","Type":"ContainerStarted","Data":"34ae9bf1fb23129c87462e808e9e5c52787007b26a7cf7d4db96e3dfb93e1394"} Oct 04 04:35:25 crc kubenswrapper[4992]: W1004 04:35:25.220662 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-b932714d1b8ef508e1ec4046a31fc22826ac57ef1463582464804ab0a4e4369e WatchSource:0}: Error finding container b932714d1b8ef508e1ec4046a31fc22826ac57ef1463582464804ab0a4e4369e: Status 404 returned error can't find the container with id b932714d1b8ef508e1ec4046a31fc22826ac57ef1463582464804ab0a4e4369e Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.221830 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" event={"ID":"0b337efa-c799-4fce-833c-cd0149c3644c","Type":"ContainerStarted","Data":"5d9cb936589e7286bd28bd0d54a9ea8247c9195a6416ff868d80ba776353cec3"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.221877 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" event={"ID":"0b337efa-c799-4fce-833c-cd0149c3644c","Type":"ContainerStarted","Data":"326ebe3f6cfe27277966860ff9584cfe875332c194e9095844410689af8e7e70"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.229282 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" event={"ID":"ee4b8c43-5789-4d06-884e-f2fcc355eeec","Type":"ContainerStarted","Data":"27ead3d3f43bdede09c3827abb6f736a3996fe21b3dbc48dfa59cd63e6e2d29a"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.229338 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" event={"ID":"ee4b8c43-5789-4d06-884e-f2fcc355eeec","Type":"ContainerStarted","Data":"f31a06270abf47b7476852f7b72a87ef19a2897f394eba9ac188538f806bbf14"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.230008 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.255626 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.256947 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.756932046 +0000 UTC m=+159.604607514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.261742 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" event={"ID":"86a9fb33-b1f8-4f50-9785-a5504e4b4efd","Type":"ContainerStarted","Data":"cf4c7c24cadae183c7a2ae23b94c81a8bdd184b9a0a78d2528be8786158ae1c7"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.261939 4992 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tz4gp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.261995 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" podUID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.262797 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.300525 4992 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rplg6 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.300606 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" podUID="86a9fb33-b1f8-4f50-9785-a5504e4b4efd" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.321154 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:25 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:25 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:25 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.321247 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.325720 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-29qfq" event={"ID":"5b11a69d-9e3a-4e77-8c17-a9f2cc34873d","Type":"ContainerStarted","Data":"9d9db22fa7a3c904926f2b97de410b199e34fef28ad7692a993c5c4bf274b3c5"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.357621 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.358689 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.85867021 +0000 UTC m=+159.706345678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.374062 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" event={"ID":"24abe58b-4b74-4207-ba8f-1b16c22617cc","Type":"ContainerStarted","Data":"dfc12adf892ce475e076b6342b9a6c6683f9436b40f71461d573eb93665c120b"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.379742 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" event={"ID":"24abe58b-4b74-4207-ba8f-1b16c22617cc","Type":"ContainerStarted","Data":"bbb3f2f79ddb48c6a537f07202b8b68e91152e82a86798e33b37ed7b51e43091"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.411491 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-78r68" event={"ID":"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3","Type":"ContainerStarted","Data":"e0432cd50d2f5793683dc6bfd329af8608347a111b28d02a25a60ad020615ea3"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.447501 4992 generic.go:334] "Generic (PLEG): container finished" podID="533ee8bc-a31e-43a7-b87c-911ecd5c191d" containerID="3efeab8799fad6d528d6aaed58ca241e0ae08988a852a41e49c1a41dacb4abd8" exitCode=0 Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.447599 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" event={"ID":"533ee8bc-a31e-43a7-b87c-911ecd5c191d","Type":"ContainerDied","Data":"3efeab8799fad6d528d6aaed58ca241e0ae08988a852a41e49c1a41dacb4abd8"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.459030 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.460445 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:25.960427823 +0000 UTC m=+159.808103291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.468138 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" event={"ID":"103e81ce-51a3-495a-adf7-87e8af781036","Type":"ContainerStarted","Data":"90b51177cc99cb244634984b09f3fe1091df8db1877da4e19317c7e636b84093"} Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.470212 4992 patch_prober.go:28] interesting pod/console-operator-58897d9998-hmx5v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.470254 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" podUID="55b54660-afa7-4a76-b86c-9d1fd3f33769" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.569445 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.570571 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.070554739 +0000 UTC m=+159.918230197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.582620 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" podStartSLOduration=129.582594139 podStartE2EDuration="2m9.582594139s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:25.567490838 +0000 UTC m=+159.415166306" watchObservedRunningTime="2025-10-04 04:35:25.582594139 +0000 UTC m=+159.430269607" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.639369 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xrgdh" podStartSLOduration=129.639279765 podStartE2EDuration="2m9.639279765s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:25.598628595 +0000 UTC m=+159.446304063" watchObservedRunningTime="2025-10-04 04:35:25.639279765 +0000 UTC m=+159.486955233" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.642473 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" podStartSLOduration=130.642447909 podStartE2EDuration="2m10.642447909s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:25.639910542 +0000 UTC m=+159.487586020" watchObservedRunningTime="2025-10-04 04:35:25.642447909 +0000 UTC m=+159.490123387" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.673458 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.673805 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.173762301 +0000 UTC m=+160.021437939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.694617 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.694699 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.716708 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" podStartSLOduration=130.716667491 podStartE2EDuration="2m10.716667491s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:25.708157045 +0000 UTC m=+159.555832523" watchObservedRunningTime="2025-10-04 04:35:25.716667491 +0000 UTC m=+159.564342979" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.781065 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.781438 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.281423332 +0000 UTC m=+160.129098800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.882704 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.883579 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.383560825 +0000 UTC m=+160.231236293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.908961 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" podStartSLOduration=129.90893985 podStartE2EDuration="2m9.90893985s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:25.843496521 +0000 UTC m=+159.691171989" watchObservedRunningTime="2025-10-04 04:35:25.90893985 +0000 UTC m=+159.756615318" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.910624 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" podStartSLOduration=129.910618574 podStartE2EDuration="2m9.910618574s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:25.908040306 +0000 UTC m=+159.755715774" watchObservedRunningTime="2025-10-04 04:35:25.910618574 +0000 UTC m=+159.758294042" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.964011 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cq9q" podStartSLOduration=129.963989002 podStartE2EDuration="2m9.963989002s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:25.962285867 +0000 UTC m=+159.809961335" watchObservedRunningTime="2025-10-04 04:35:25.963989002 +0000 UTC m=+159.811664470" Oct 04 04:35:25 crc kubenswrapper[4992]: I1004 04:35:25.984865 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:25 crc kubenswrapper[4992]: E1004 04:35:25.985292 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.485278628 +0000 UTC m=+160.332954096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.006763 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" podStartSLOduration=130.006746338 podStartE2EDuration="2m10.006746338s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.004214601 +0000 UTC m=+159.851890069" watchObservedRunningTime="2025-10-04 04:35:26.006746338 +0000 UTC m=+159.854421806" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.061776 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-r7t8c" podStartSLOduration=130.06174287 podStartE2EDuration="2m10.06174287s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.061281547 +0000 UTC m=+159.908957025" watchObservedRunningTime="2025-10-04 04:35:26.06174287 +0000 UTC m=+159.909418338" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.062262 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" podStartSLOduration=130.062256043 podStartE2EDuration="2m10.062256043s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.033610432 +0000 UTC m=+159.881285900" watchObservedRunningTime="2025-10-04 04:35:26.062256043 +0000 UTC m=+159.909931511" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.086509 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.086989 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.5869727 +0000 UTC m=+160.434648168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.121548 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" podStartSLOduration=130.121529058 podStartE2EDuration="2m10.121529058s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.117311786 +0000 UTC m=+159.964987254" watchObservedRunningTime="2025-10-04 04:35:26.121529058 +0000 UTC m=+159.969204516" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.155175 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5sgbq" podStartSLOduration=130.155138801 podStartE2EDuration="2m10.155138801s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.152161652 +0000 UTC m=+159.999837120" watchObservedRunningTime="2025-10-04 04:35:26.155138801 +0000 UTC m=+160.002814269" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.189204 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.189572 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.689558325 +0000 UTC m=+160.537233793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.210152 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" podStartSLOduration=130.210121142 podStartE2EDuration="2m10.210121142s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.207392609 +0000 UTC m=+160.055068077" watchObservedRunningTime="2025-10-04 04:35:26.210121142 +0000 UTC m=+160.057796610" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.215319 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-78r68" podStartSLOduration=130.215300839 podStartE2EDuration="2m10.215300839s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.178030359 +0000 UTC m=+160.025705827" watchObservedRunningTime="2025-10-04 04:35:26.215300839 +0000 UTC m=+160.062976307" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.291373 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.291616 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:26 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:26 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:26 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.291670 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.291861 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.791842673 +0000 UTC m=+160.639518141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.293109 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-b79wf" podStartSLOduration=130.293085146 podStartE2EDuration="2m10.293085146s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.277464871 +0000 UTC m=+160.125140339" watchObservedRunningTime="2025-10-04 04:35:26.293085146 +0000 UTC m=+160.140760614" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.324968 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" podStartSLOduration=131.324938182 podStartE2EDuration="2m11.324938182s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.322655202 +0000 UTC m=+160.170330670" watchObservedRunningTime="2025-10-04 04:35:26.324938182 +0000 UTC m=+160.172613650" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.348176 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-29qfq" podStartSLOduration=8.348160669 podStartE2EDuration="8.348160669s" podCreationTimestamp="2025-10-04 04:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.348091168 +0000 UTC m=+160.195766636" watchObservedRunningTime="2025-10-04 04:35:26.348160669 +0000 UTC m=+160.195836137" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.376777 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n9lrz" podStartSLOduration=130.376759829 podStartE2EDuration="2m10.376759829s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.37378821 +0000 UTC m=+160.221463678" watchObservedRunningTime="2025-10-04 04:35:26.376759829 +0000 UTC m=+160.224435297" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.395022 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.395622 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.89561014 +0000 UTC m=+160.743285608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.499282 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.499711 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:26.999696026 +0000 UTC m=+160.847371494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.517277 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" event={"ID":"b209aa10-3135-455f-8ad0-a54f64b7768d","Type":"ContainerStarted","Data":"4818b4a35bab641223a77488dbb01c6a61091f45b92237ef4bf58c82f1f9bf49"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.517327 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" event={"ID":"b209aa10-3135-455f-8ad0-a54f64b7768d","Type":"ContainerStarted","Data":"c23b587536d2ed2bdc3ba3492a739b8aef2d9f20a40bcf827b2e700867bea4b6"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.550098 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"462f423f58e17bf3a55cf62652e0174fc7be0b40464c3f3b68cf64994c5accaf"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.550144 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7bbbf9337b19396ed64ac09497d85e7b5f55eab563278682de9a9b5ee9b07ae4"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.550410 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.568751 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" event={"ID":"4256fad3-f7d5-4441-a4f0-ee84712d7321","Type":"ContainerStarted","Data":"9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.572783 4992 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4tqf8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.572835 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" podUID="4256fad3-f7d5-4441-a4f0-ee84712d7321" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.575032 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-frs8b" podStartSLOduration=130.575009257 podStartE2EDuration="2m10.575009257s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.573562168 +0000 UTC m=+160.421237636" watchObservedRunningTime="2025-10-04 04:35:26.575009257 +0000 UTC m=+160.422684725" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.601665 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.603265 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.103253517 +0000 UTC m=+160.950928975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.610221 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4f9mb" event={"ID":"9d143445-9a26-4203-995d-82238aa12f54","Type":"ContainerStarted","Data":"2ebdc7168b8cbd1c9a4bbeeb0bc280f253fbde99b5edd08aeb7022a011a69fb4"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.635893 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-78r68" event={"ID":"c5e3a93f-4aa4-4d6d-b144-3a65b33a2eb3","Type":"ContainerStarted","Data":"e7ec55db6c672b151ab49bebc0ef681d07a88acfca47452e8459e8794cff4f74"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.637875 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fwb7s" event={"ID":"78b0d5e3-fbae-4a2b-9ddf-6c8cdbdb319d","Type":"ContainerStarted","Data":"b3cd5ee96ef39c52922f16ea674f8983dbfd97fabf95b2b8f35265ef539bca06"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.639341 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-blzx5" event={"ID":"7206a358-c3d3-4f1b-ac51-4c0657f9e29a","Type":"ContainerStarted","Data":"94fcc02e52dfabf649b95f98fb06e45833be2248cf366fdbc74a272692a80c2a"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.639918 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.640922 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" event={"ID":"0b337efa-c799-4fce-833c-cd0149c3644c","Type":"ContainerStarted","Data":"e0c4d1deac6ac686057d6dc5195002c71159b980a2e3212df4b3cf1929bae7ff"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.672315 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"35fdfcaa78081deb0c3dd3ef8d99e5f3f4c7937a7d38a75c4bcb1fbede737a71"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.672404 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b932714d1b8ef508e1ec4046a31fc22826ac57ef1463582464804ab0a4e4369e"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.679644 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" event={"ID":"20f97015-5d3a-43a3-8863-95fcc8dd8414","Type":"ContainerStarted","Data":"6af856df338274e28e50c76eef4e3c05f0c87ee5246d4fe685ee41fe633118c0"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.701002 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-29qfq" event={"ID":"5b11a69d-9e3a-4e77-8c17-a9f2cc34873d","Type":"ContainerStarted","Data":"59f790f908266afa6d367481f8c475399f855f6156a8c362fe3f215e3852126c"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.702409 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.703237 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.203222443 +0000 UTC m=+161.050897911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.724440 4992 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-m8ccp container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 04:35:26 crc kubenswrapper[4992]: [+]log ok Oct 04 04:35:26 crc kubenswrapper[4992]: [+]etcd ok Oct 04 04:35:26 crc kubenswrapper[4992]: [-]poststarthook/start-apiserver-admission-initializer failed: reason withheld Oct 04 04:35:26 crc kubenswrapper[4992]: [-]poststarthook/generic-apiserver-start-informers failed: reason withheld Oct 04 04:35:26 crc kubenswrapper[4992]: [-]poststarthook/max-in-flight-filter failed: reason withheld Oct 04 04:35:26 crc kubenswrapper[4992]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 04:35:26 crc kubenswrapper[4992]: [+]poststarthook/openshift.io-StartUserInformer ok Oct 04 04:35:26 crc kubenswrapper[4992]: [+]poststarthook/openshift.io-StartOAuthInformer ok Oct 04 04:35:26 crc kubenswrapper[4992]: [-]poststarthook/openshift.io-StartTokenTimeoutUpdater failed: reason withheld Oct 04 04:35:26 crc kubenswrapper[4992]: livez check failed Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.724495 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" podUID="3cd8a6e8-ee96-4e7e-9367-e74478678f5f" containerName="oauth-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.725402 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" event={"ID":"35350b4a-c3a6-4d70-bab2-a79757701029","Type":"ContainerStarted","Data":"d5d2e8b76910136be1e92e9cae8171a71d932944370ff492e98fc79b39c4ae4e"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.744398 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e0b2bec4ab1cd951e6e7c22cf0819e182d6eb14689c68c7f9f52345c8fd8b004"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.744446 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9699769b20f237ad254de18ac80c5ccdf93bc3010ce0f2bae0aac93dddb12b65"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.772784 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-blzx5" podStartSLOduration=8.772764161 podStartE2EDuration="8.772764161s" podCreationTimestamp="2025-10-04 04:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.706390607 +0000 UTC m=+160.554066075" watchObservedRunningTime="2025-10-04 04:35:26.772764161 +0000 UTC m=+160.620439629" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.776691 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-x8sls" podStartSLOduration=130.776674115 podStartE2EDuration="2m10.776674115s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.771804015 +0000 UTC m=+160.619479483" watchObservedRunningTime="2025-10-04 04:35:26.776674115 +0000 UTC m=+160.624349583" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.777277 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" event={"ID":"0784a997-5d78-4199-a3fa-3360eef7a756","Type":"ContainerStarted","Data":"aeddbcc94e2382cd1f2a0ba461ebf0ab72155eab1ae584e53cf861597d6dfa62"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.777921 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.803991 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.804431 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.304419222 +0000 UTC m=+161.152094690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.806981 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" event={"ID":"533ee8bc-a31e-43a7-b87c-911ecd5c191d","Type":"ContainerStarted","Data":"d72ff8fa1147e25855f5227ad4e912b0a48698b329d948bad51c95e71aeec077"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.827582 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-b99bg" podStartSLOduration=130.827563437 podStartE2EDuration="2m10.827563437s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:26.823575241 +0000 UTC m=+160.671250709" watchObservedRunningTime="2025-10-04 04:35:26.827563437 +0000 UTC m=+160.675238905" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.861586 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" event={"ID":"936b65e0-cec1-4d1f-a7ce-2f3cc142a9c8","Type":"ContainerStarted","Data":"ddadc1c1262e6af64a2d09024a7ba54a335ac47faa7cfc5e9983270f255a2826"} Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.862398 4992 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tz4gp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.862448 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" podUID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.863283 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.863307 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.901119 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z2mg2" Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.912984 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:26 crc kubenswrapper[4992]: E1004 04:35:26.926601 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.426581388 +0000 UTC m=+161.274256856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:26 crc kubenswrapper[4992]: I1004 04:35:26.983304 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rplg6" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.005416 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-79tvb" podStartSLOduration=131.005375681 podStartE2EDuration="2m11.005375681s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:27.003421789 +0000 UTC m=+160.851097257" watchObservedRunningTime="2025-10-04 04:35:27.005375681 +0000 UTC m=+160.853051159" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.019147 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.019533 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.519520077 +0000 UTC m=+161.367195545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.118508 4992 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-ftjbd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.118885 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" podUID="82d8b4e3-0fe2-4941-895e-47513a81b9bf" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.121087 4992 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-ftjbd container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.121123 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" podUID="82d8b4e3-0fe2-4941-895e-47513a81b9bf" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.123008 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.123405 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.623386736 +0000 UTC m=+161.471062204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.185272 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k5knb" podStartSLOduration=131.18525379 podStartE2EDuration="2m11.18525379s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:27.18524555 +0000 UTC m=+161.032921028" watchObservedRunningTime="2025-10-04 04:35:27.18525379 +0000 UTC m=+161.032929258" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.224527 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.224854 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.724841552 +0000 UTC m=+161.572517020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.249544 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-hmx5v" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.287426 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:27 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:27 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:27 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.287503 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.309005 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" podStartSLOduration=132.308984397 podStartE2EDuration="2m12.308984397s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:27.307525108 +0000 UTC m=+161.155200586" watchObservedRunningTime="2025-10-04 04:35:27.308984397 +0000 UTC m=+161.156659875" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.325683 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.325911 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.825880376 +0000 UTC m=+161.673555844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.325975 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.326410 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.826390219 +0000 UTC m=+161.674065867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.427038 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.427231 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.927202518 +0000 UTC m=+161.774877986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.427346 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.427750 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:27.927738382 +0000 UTC m=+161.775413850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.430258 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" podStartSLOduration=131.430238578 podStartE2EDuration="2m11.430238578s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:27.400911159 +0000 UTC m=+161.248586647" watchObservedRunningTime="2025-10-04 04:35:27.430238578 +0000 UTC m=+161.277914046" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.528970 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.529186 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.029152096 +0000 UTC m=+161.876827584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.529283 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.529724 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.029708121 +0000 UTC m=+161.877383749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.630063 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.630259 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.130229702 +0000 UTC m=+161.977905160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.630398 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.630710 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.130696574 +0000 UTC m=+161.978372032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.731632 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.731852 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.231817321 +0000 UTC m=+162.079492789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.731937 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.732328 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.232315974 +0000 UTC m=+162.079991622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.833523 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.833739 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.333708758 +0000 UTC m=+162.181384216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.833882 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.834237 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.334228072 +0000 UTC m=+162.181903540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.861445 4992 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-rd62r container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.861525 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" podUID="dd501849-1d14-48c1-bc51-06d23c56db28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.862372 4992 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-xjmbw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.862441 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" podUID="aee1200b-878a-4a90-a34c-dcc237c6a8a3" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.881426 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" event={"ID":"533ee8bc-a31e-43a7-b87c-911ecd5c191d","Type":"ContainerStarted","Data":"1fcd0afea66c73e57ed12d9444d5a0fa354079805765a7d4e2bee4ec6e5eb47d"} Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.886993 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" event={"ID":"617d3d62-5b4e-481b-817c-54288406fc7c","Type":"ContainerStarted","Data":"e1809a2bafe3408b0d22dc76207447cf86c69cba356ccf61b3ac9de39f1e8714"} Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.890161 4992 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tz4gp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" start-of-body= Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.890203 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" podUID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.43:8080/healthz\": dial tcp 10.217.0.43:8080: connect: connection refused" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.893538 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.893615 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.897429 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:35:27 crc kubenswrapper[4992]: I1004 04:35:27.934920 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:27 crc kubenswrapper[4992]: E1004 04:35:27.936461 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.436427247 +0000 UTC m=+162.284102825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.037732 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.038156 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.53813976 +0000 UTC m=+162.385815228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.139329 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.139541 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.639513813 +0000 UTC m=+162.487189281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.139587 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.139990 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.639975385 +0000 UTC m=+162.487650853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.243068 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.243243 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.743208868 +0000 UTC m=+162.590884346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.243290 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.243626 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.743615309 +0000 UTC m=+162.591290777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.293904 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:28 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:28 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:28 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.293975 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.344676 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.344890 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.844856469 +0000 UTC m=+162.692531937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.344956 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.346038 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.846005779 +0000 UTC m=+162.693681427 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.348152 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.446750 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.446866 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.946847729 +0000 UTC m=+162.794523197 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.447052 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.447491 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:28.947479315 +0000 UTC m=+162.795154783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.508735 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c4mz6"] Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.511305 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.526295 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.528718 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4mz6"] Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.549996 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.550317 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.050287017 +0000 UTC m=+162.897962485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.550406 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-utilities\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.550486 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfgx4\" (UniqueName: \"kubernetes.io/projected/cba17782-3829-4de8-b49e-9eba858fc1ed-kube-api-access-tfgx4\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.550550 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.550615 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-catalog-content\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.550972 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.050962755 +0000 UTC m=+162.898638223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.651499 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.651674 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.15164993 +0000 UTC m=+162.999325408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.651750 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-utilities\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.651819 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfgx4\" (UniqueName: \"kubernetes.io/projected/cba17782-3829-4de8-b49e-9eba858fc1ed-kube-api-access-tfgx4\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.651865 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.651911 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-catalog-content\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.652385 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-catalog-content\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.652475 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.152463562 +0000 UTC m=+163.000139030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.652924 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-utilities\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.661918 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g8cmz"] Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.663152 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.665049 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.675811 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g8cmz"] Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.707918 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfgx4\" (UniqueName: \"kubernetes.io/projected/cba17782-3829-4de8-b49e-9eba858fc1ed-kube-api-access-tfgx4\") pod \"certified-operators-c4mz6\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.754305 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.754700 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-utilities\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.754747 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ld6g\" (UniqueName: \"kubernetes.io/projected/35ec2da1-7bca-4056-96eb-67a9687c4472-kube-api-access-2ld6g\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.754794 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-catalog-content\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.755003 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.254981346 +0000 UTC m=+163.102656814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.794474 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ftjbd" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.850192 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.856223 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-utilities\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.856486 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ld6g\" (UniqueName: \"kubernetes.io/projected/35ec2da1-7bca-4056-96eb-67a9687c4472-kube-api-access-2ld6g\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.856641 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-catalog-content\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.856789 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.856961 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-utilities\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.857337 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.357322575 +0000 UTC m=+163.204998043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.857510 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-catalog-content\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.864472 4992 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.885921 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vwqln"] Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.887494 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.888466 4992 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-xjmbw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.888512 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" podUID="aee1200b-878a-4a90-a34c-dcc237c6a8a3" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.889066 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ld6g\" (UniqueName: \"kubernetes.io/projected/35ec2da1-7bca-4056-96eb-67a9687c4472-kube-api-access-2ld6g\") pod \"community-operators-g8cmz\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.905747 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vwqln"] Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.917961 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.922739 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.925456 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.928079 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.945784 4992 generic.go:334] "Generic (PLEG): container finished" podID="24abe58b-4b74-4207-ba8f-1b16c22617cc" containerID="dfc12adf892ce475e076b6342b9a6c6683f9436b40f71461d573eb93665c120b" exitCode=0 Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.945881 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" event={"ID":"24abe58b-4b74-4207-ba8f-1b16c22617cc","Type":"ContainerDied","Data":"dfc12adf892ce475e076b6342b9a6c6683f9436b40f71461d573eb93665c120b"} Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.957611 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.957722 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.457706402 +0000 UTC m=+163.305381870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.957871 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e72c23e9-d135-4282-801c-c6f4bddd318b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e72c23e9-d135-4282-801c-c6f4bddd318b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.957916 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pg4m\" (UniqueName: \"kubernetes.io/projected/1463c32f-0eba-42da-b876-5c714cdf3c88-kube-api-access-4pg4m\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.957960 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e72c23e9-d135-4282-801c-c6f4bddd318b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e72c23e9-d135-4282-801c-c6f4bddd318b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.958018 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-catalog-content\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.958045 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.958060 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-utilities\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:28 crc kubenswrapper[4992]: E1004 04:35:28.958306 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.458299068 +0000 UTC m=+163.305974536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.959305 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 04:35:28 crc kubenswrapper[4992]: I1004 04:35:28.994192 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.042752 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" event={"ID":"617d3d62-5b4e-481b-817c-54288406fc7c","Type":"ContainerStarted","Data":"25d6c5a80d12d5d75519dea5daf1141a828c1a3355155b7e14e774d16d8066ed"} Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.042788 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" event={"ID":"617d3d62-5b4e-481b-817c-54288406fc7c","Type":"ContainerStarted","Data":"5dcc1b4f826c99865aa1baa9a8561d29b3c35d5582462e6d193bd6ceedfed4a4"} Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.061940 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.062109 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-catalog-content\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.062140 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-utilities\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.062180 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e72c23e9-d135-4282-801c-c6f4bddd318b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e72c23e9-d135-4282-801c-c6f4bddd318b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.062203 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pg4m\" (UniqueName: \"kubernetes.io/projected/1463c32f-0eba-42da-b876-5c714cdf3c88-kube-api-access-4pg4m\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.062225 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e72c23e9-d135-4282-801c-c6f4bddd318b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e72c23e9-d135-4282-801c-c6f4bddd318b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.062308 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e72c23e9-d135-4282-801c-c6f4bddd318b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e72c23e9-d135-4282-801c-c6f4bddd318b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:29 crc kubenswrapper[4992]: E1004 04:35:29.062400 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.562370853 +0000 UTC m=+163.410046321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.063080 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-catalog-content\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.063103 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-utilities\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.078123 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wp4ff"] Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.086793 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pg4m\" (UniqueName: \"kubernetes.io/projected/1463c32f-0eba-42da-b876-5c714cdf3c88-kube-api-access-4pg4m\") pod \"certified-operators-vwqln\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.089433 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.093989 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e72c23e9-d135-4282-801c-c6f4bddd318b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e72c23e9-d135-4282-801c-c6f4bddd318b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.117869 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wp4ff"] Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.163167 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-utilities\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.163472 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvhkk\" (UniqueName: \"kubernetes.io/projected/5e7c9252-8288-416d-a5bb-0e1ffb586240-kube-api-access-nvhkk\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.163640 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-catalog-content\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.163780 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:29 crc kubenswrapper[4992]: E1004 04:35:29.164061 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.664049354 +0000 UTC m=+163.511724822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.304799 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:29 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:29 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:29 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.304900 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.306665 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.308083 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.308300 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-catalog-content\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.308370 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-utilities\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.309187 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:29 crc kubenswrapper[4992]: E1004 04:35:29.309315 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.809298253 +0000 UTC m=+163.656973721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.309911 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvhkk\" (UniqueName: \"kubernetes.io/projected/5e7c9252-8288-416d-a5bb-0e1ffb586240-kube-api-access-nvhkk\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.309969 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-catalog-content\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.310278 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-utilities\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.339922 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvhkk\" (UniqueName: \"kubernetes.io/projected/5e7c9252-8288-416d-a5bb-0e1ffb586240-kube-api-access-nvhkk\") pod \"community-operators-wp4ff\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.413427 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:29 crc kubenswrapper[4992]: E1004 04:35:29.413931 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 04:35:29.913908473 +0000 UTC m=+163.761584111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fz68t" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.479466 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.515905 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:29 crc kubenswrapper[4992]: E1004 04:35:29.516349 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 04:35:30.016328864 +0000 UTC m=+163.864004332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.573764 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4mz6"] Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.574659 4992 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T04:35:28.864510346Z","Handler":null,"Name":""} Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.585707 4992 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.585738 4992 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.588519 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vwqln"] Oct 04 04:35:29 crc kubenswrapper[4992]: W1004 04:35:29.594661 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1463c32f_0eba_42da_b876_5c714cdf3c88.slice/crio-4eca2325aa5397381a82ea7fc6b1e93e5a537b6a783420b792ba7919bcd6d449 WatchSource:0}: Error finding container 4eca2325aa5397381a82ea7fc6b1e93e5a537b6a783420b792ba7919bcd6d449: Status 404 returned error can't find the container with id 4eca2325aa5397381a82ea7fc6b1e93e5a537b6a783420b792ba7919bcd6d449 Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.618156 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.648938 4992 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.649013 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.720508 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fz68t\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.743698 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g8cmz"] Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.814829 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wp4ff"] Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.822929 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.827618 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.868857 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 04:35:29 crc kubenswrapper[4992]: W1004 04:35:29.959845 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode72c23e9_d135_4282_801c_c6f4bddd318b.slice/crio-e2261788e2f7cff0b430cd3bc4ef3cec4e71baacbc2b5d529f3310881a6fa6ae WatchSource:0}: Error finding container e2261788e2f7cff0b430cd3bc4ef3cec4e71baacbc2b5d529f3310881a6fa6ae: Status 404 returned error can't find the container with id e2261788e2f7cff0b430cd3bc4ef3cec4e71baacbc2b5d529f3310881a6fa6ae Oct 04 04:35:29 crc kubenswrapper[4992]: I1004 04:35:29.999886 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.054639 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.055492 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.057912 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e72c23e9-d135-4282-801c-c6f4bddd318b","Type":"ContainerStarted","Data":"e2261788e2f7cff0b430cd3bc4ef3cec4e71baacbc2b5d529f3310881a6fa6ae"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.060350 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.060703 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.065983 4992 generic.go:334] "Generic (PLEG): container finished" podID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerID="ea07cd786f44516980c9679d4dbd0f08f692235b18a98a71e5312fc6849aceab" exitCode=0 Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.066064 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8cmz" event={"ID":"35ec2da1-7bca-4056-96eb-67a9687c4472","Type":"ContainerDied","Data":"ea07cd786f44516980c9679d4dbd0f08f692235b18a98a71e5312fc6849aceab"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.066098 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8cmz" event={"ID":"35ec2da1-7bca-4056-96eb-67a9687c4472","Type":"ContainerStarted","Data":"d8f66a612b2340b88f682d55763868b9933ed17f8a1f2b6f67592ea946ddac56"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.069851 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.073339 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.083530 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" event={"ID":"617d3d62-5b4e-481b-817c-54288406fc7c","Type":"ContainerStarted","Data":"0facfa7ff0261236683c4656c922489ef6520dd774184776c6d07c20e870afd4"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.090813 4992 generic.go:334] "Generic (PLEG): container finished" podID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerID="2f63cbc106580d36e0c6f107171833104935d71765c8f80b75f6b14b987b610c" exitCode=0 Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.090896 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mz6" event={"ID":"cba17782-3829-4de8-b49e-9eba858fc1ed","Type":"ContainerDied","Data":"2f63cbc106580d36e0c6f107171833104935d71765c8f80b75f6b14b987b610c"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.090928 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mz6" event={"ID":"cba17782-3829-4de8-b49e-9eba858fc1ed","Type":"ContainerStarted","Data":"fcc012b1f36bef5c5a4540ec3120247695a361208cec40347cad6f81bc5c0625"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.102594 4992 generic.go:334] "Generic (PLEG): container finished" podID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerID="81845d871d53bdbdbeacb241de7c3be39e959b7a1cc4b8d04c92b15b691bed9b" exitCode=0 Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.103384 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwqln" event={"ID":"1463c32f-0eba-42da-b876-5c714cdf3c88","Type":"ContainerDied","Data":"81845d871d53bdbdbeacb241de7c3be39e959b7a1cc4b8d04c92b15b691bed9b"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.103436 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwqln" event={"ID":"1463c32f-0eba-42da-b876-5c714cdf3c88","Type":"ContainerStarted","Data":"4eca2325aa5397381a82ea7fc6b1e93e5a537b6a783420b792ba7919bcd6d449"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.112524 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4ff" event={"ID":"5e7c9252-8288-416d-a5bb-0e1ffb586240","Type":"ContainerStarted","Data":"22e690f0338c1aeb3a8ef2408a67987e33b835d97f9534de85a9922406b53389"} Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.129749 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0526e03e-5889-4202-a969-e2fec17fca8b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0526e03e-5889-4202-a969-e2fec17fca8b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.130995 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0526e03e-5889-4202-a969-e2fec17fca8b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0526e03e-5889-4202-a969-e2fec17fca8b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.182251 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-ttqhz" podStartSLOduration=12.182224546 podStartE2EDuration="12.182224546s" podCreationTimestamp="2025-10-04 04:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:30.153948775 +0000 UTC m=+164.001624243" watchObservedRunningTime="2025-10-04 04:35:30.182224546 +0000 UTC m=+164.029900014" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.232673 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0526e03e-5889-4202-a969-e2fec17fca8b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0526e03e-5889-4202-a969-e2fec17fca8b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.232715 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0526e03e-5889-4202-a969-e2fec17fca8b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0526e03e-5889-4202-a969-e2fec17fca8b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.234500 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0526e03e-5889-4202-a969-e2fec17fca8b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0526e03e-5889-4202-a969-e2fec17fca8b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.262371 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fz68t"] Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.268298 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0526e03e-5889-4202-a969-e2fec17fca8b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0526e03e-5889-4202-a969-e2fec17fca8b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.294479 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:30 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:30 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:30 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.294526 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.360130 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.384662 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.449745 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.537336 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm5lq\" (UniqueName: \"kubernetes.io/projected/24abe58b-4b74-4207-ba8f-1b16c22617cc-kube-api-access-bm5lq\") pod \"24abe58b-4b74-4207-ba8f-1b16c22617cc\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.537438 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24abe58b-4b74-4207-ba8f-1b16c22617cc-secret-volume\") pod \"24abe58b-4b74-4207-ba8f-1b16c22617cc\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.537476 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24abe58b-4b74-4207-ba8f-1b16c22617cc-config-volume\") pod \"24abe58b-4b74-4207-ba8f-1b16c22617cc\" (UID: \"24abe58b-4b74-4207-ba8f-1b16c22617cc\") " Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.538292 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24abe58b-4b74-4207-ba8f-1b16c22617cc-config-volume" (OuterVolumeSpecName: "config-volume") pod "24abe58b-4b74-4207-ba8f-1b16c22617cc" (UID: "24abe58b-4b74-4207-ba8f-1b16c22617cc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.547872 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24abe58b-4b74-4207-ba8f-1b16c22617cc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "24abe58b-4b74-4207-ba8f-1b16c22617cc" (UID: "24abe58b-4b74-4207-ba8f-1b16c22617cc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.548002 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24abe58b-4b74-4207-ba8f-1b16c22617cc-kube-api-access-bm5lq" (OuterVolumeSpecName: "kube-api-access-bm5lq") pod "24abe58b-4b74-4207-ba8f-1b16c22617cc" (UID: "24abe58b-4b74-4207-ba8f-1b16c22617cc"). InnerVolumeSpecName "kube-api-access-bm5lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.638572 4992 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/24abe58b-4b74-4207-ba8f-1b16c22617cc-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.638607 4992 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/24abe58b-4b74-4207-ba8f-1b16c22617cc-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.638619 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm5lq\" (UniqueName: \"kubernetes.io/projected/24abe58b-4b74-4207-ba8f-1b16c22617cc-kube-api-access-bm5lq\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.653443 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w57r2"] Oct 04 04:35:30 crc kubenswrapper[4992]: E1004 04:35:30.653674 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24abe58b-4b74-4207-ba8f-1b16c22617cc" containerName="collect-profiles" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.653685 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="24abe58b-4b74-4207-ba8f-1b16c22617cc" containerName="collect-profiles" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.653788 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="24abe58b-4b74-4207-ba8f-1b16c22617cc" containerName="collect-profiles" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.654514 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.660719 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.663950 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.666760 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w57r2"] Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.703387 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.716030 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m8ccp" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.739963 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82qwm\" (UniqueName: \"kubernetes.io/projected/151de3a2-278d-40d7-b869-023e9cd39273-kube-api-access-82qwm\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.740045 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-catalog-content\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.740084 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-utilities\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.758795 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.841772 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82qwm\" (UniqueName: \"kubernetes.io/projected/151de3a2-278d-40d7-b869-023e9cd39273-kube-api-access-82qwm\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.841827 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-catalog-content\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.841847 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-utilities\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.842301 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-utilities\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.842958 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-catalog-content\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.872770 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82qwm\" (UniqueName: \"kubernetes.io/projected/151de3a2-278d-40d7-b869-023e9cd39273-kube-api-access-82qwm\") pod \"redhat-marketplace-w57r2\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:30 crc kubenswrapper[4992]: I1004 04:35:30.971781 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.067401 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dpzhh"] Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.068663 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.078968 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpzhh"] Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.123576 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e72c23e9-d135-4282-801c-c6f4bddd318b","Type":"ContainerStarted","Data":"6e1247293aa2378c1dd28ede9385d1efd8c5f403b8407513ae79ca6970bd4860"} Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.127670 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" event={"ID":"78372347-a30d-4a0f-a4d4-c92a1999eaad","Type":"ContainerStarted","Data":"de510d584b716716a3dbc0eb8972939ce14fb13130706471106edce880eca207"} Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.127714 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" event={"ID":"78372347-a30d-4a0f-a4d4-c92a1999eaad","Type":"ContainerStarted","Data":"0751b6b8404873485d3c9ce2df880f8943e6fb368d74adf09a839b8d8f4ca9fb"} Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.128626 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.141056 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0526e03e-5889-4202-a969-e2fec17fca8b","Type":"ContainerStarted","Data":"d192b94227d2cac6e297703f64e1a1efbc181b7c676657566b0581df892e2d28"} Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.145336 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.145313033 podStartE2EDuration="3.145313033s" podCreationTimestamp="2025-10-04 04:35:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:31.143953107 +0000 UTC m=+164.991628585" watchObservedRunningTime="2025-10-04 04:35:31.145313033 +0000 UTC m=+164.992988501" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.161062 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" event={"ID":"24abe58b-4b74-4207-ba8f-1b16c22617cc","Type":"ContainerDied","Data":"bbb3f2f79ddb48c6a537f07202b8b68e91152e82a86798e33b37ed7b51e43091"} Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.161126 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbb3f2f79ddb48c6a537f07202b8b68e91152e82a86798e33b37ed7b51e43091" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.161017 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.168410 4992 generic.go:334] "Generic (PLEG): container finished" podID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerID="8fb474035d2fe0efc33beb530f362487049611706ae632e06080f51e7d8e8824" exitCode=0 Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.169597 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4ff" event={"ID":"5e7c9252-8288-416d-a5bb-0e1ffb586240","Type":"ContainerDied","Data":"8fb474035d2fe0efc33beb530f362487049611706ae632e06080f51e7d8e8824"} Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.172784 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" podStartSLOduration=135.172758053 podStartE2EDuration="2m15.172758053s" podCreationTimestamp="2025-10-04 04:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:35:31.170225285 +0000 UTC m=+165.017900763" watchObservedRunningTime="2025-10-04 04:35:31.172758053 +0000 UTC m=+165.020433521" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.177541 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.177776 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.179575 4992 patch_prober.go:28] interesting pod/console-f9d7485db-5sgbq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.179591 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.179655 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5sgbq" podUID="8e339a50-fbf4-44da-9335-ba6ae033fc38" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.179671 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.179587 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.180094 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.210054 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.212215 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.244335 4992 patch_prober.go:28] interesting pod/apiserver-76f77b778f-pjtjn container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]log ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]etcd ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/max-in-flight-filter ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 04 04:35:31 crc kubenswrapper[4992]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 04 04:35:31 crc kubenswrapper[4992]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/project.openshift.io-projectcache ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/openshift.io-startinformers ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 04 04:35:31 crc kubenswrapper[4992]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 04:35:31 crc kubenswrapper[4992]: livez check failed Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.244410 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" podUID="533ee8bc-a31e-43a7-b87c-911ecd5c191d" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.247825 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8tpx\" (UniqueName: \"kubernetes.io/projected/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-kube-api-access-x8tpx\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.247942 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-catalog-content\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.248027 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-utilities\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.287507 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.290617 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:31 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:31 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:31 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.290662 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.294109 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w57r2"] Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.348814 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8tpx\" (UniqueName: \"kubernetes.io/projected/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-kube-api-access-x8tpx\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.349251 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-catalog-content\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.349406 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-utilities\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.350034 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-xjmbw" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.352780 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-catalog-content\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.356803 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-utilities\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.387236 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8tpx\" (UniqueName: \"kubernetes.io/projected/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-kube-api-access-x8tpx\") pod \"redhat-marketplace-dpzhh\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.387556 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.683397 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-grcdz"] Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.685500 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.699611 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.709854 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-grcdz"] Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.855991 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpzhh"] Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.861559 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khdwl\" (UniqueName: \"kubernetes.io/projected/acebc5f9-1793-4b93-a1e9-55604bb69996-kube-api-access-khdwl\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.861613 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-catalog-content\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.861656 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-utilities\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: W1004 04:35:31.867768 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ea2c4fc_4545_4f6d_85f9_20a8d9ca87f1.slice/crio-15bd86efac51b7329ad1325fdf10f192841b09b107e45bb70543645d244126ca WatchSource:0}: Error finding container 15bd86efac51b7329ad1325fdf10f192841b09b107e45bb70543645d244126ca: Status 404 returned error can't find the container with id 15bd86efac51b7329ad1325fdf10f192841b09b107e45bb70543645d244126ca Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.963221 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-catalog-content\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.963296 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-utilities\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.963389 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khdwl\" (UniqueName: \"kubernetes.io/projected/acebc5f9-1793-4b93-a1e9-55604bb69996-kube-api-access-khdwl\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.964062 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-utilities\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.965409 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-catalog-content\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.981629 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:35:31 crc kubenswrapper[4992]: I1004 04:35:31.984045 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khdwl\" (UniqueName: \"kubernetes.io/projected/acebc5f9-1793-4b93-a1e9-55604bb69996-kube-api-access-khdwl\") pod \"redhat-operators-grcdz\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.021843 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.058921 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lvwbd"] Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.060495 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.070957 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lvwbd"] Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.166770 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-catalog-content\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.166845 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r42kx\" (UniqueName: \"kubernetes.io/projected/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-kube-api-access-r42kx\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.166988 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-utilities\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.183359 4992 generic.go:334] "Generic (PLEG): container finished" podID="0526e03e-5889-4202-a969-e2fec17fca8b" containerID="9b30f378a61b36b958cd41fb54df9ef243491f6f832c6e0175d44129476b2fb3" exitCode=0 Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.183484 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0526e03e-5889-4202-a969-e2fec17fca8b","Type":"ContainerDied","Data":"9b30f378a61b36b958cd41fb54df9ef243491f6f832c6e0175d44129476b2fb3"} Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.187790 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpzhh" event={"ID":"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1","Type":"ContainerStarted","Data":"fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2"} Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.187842 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpzhh" event={"ID":"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1","Type":"ContainerStarted","Data":"15bd86efac51b7329ad1325fdf10f192841b09b107e45bb70543645d244126ca"} Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.198750 4992 generic.go:334] "Generic (PLEG): container finished" podID="e72c23e9-d135-4282-801c-c6f4bddd318b" containerID="6e1247293aa2378c1dd28ede9385d1efd8c5f403b8407513ae79ca6970bd4860" exitCode=0 Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.198900 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e72c23e9-d135-4282-801c-c6f4bddd318b","Type":"ContainerDied","Data":"6e1247293aa2378c1dd28ede9385d1efd8c5f403b8407513ae79ca6970bd4860"} Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.214113 4992 generic.go:334] "Generic (PLEG): container finished" podID="151de3a2-278d-40d7-b869-023e9cd39273" containerID="0bebeb8ce526839ccafa25c11e8d56b8d8ff2f443b1b79dbe684dc7ec0a24801" exitCode=0 Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.215239 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w57r2" event={"ID":"151de3a2-278d-40d7-b869-023e9cd39273","Type":"ContainerDied","Data":"0bebeb8ce526839ccafa25c11e8d56b8d8ff2f443b1b79dbe684dc7ec0a24801"} Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.215270 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w57r2" event={"ID":"151de3a2-278d-40d7-b869-023e9cd39273","Type":"ContainerStarted","Data":"69f3de7cbb35e9e7f8ae2793e18a4f75bd2cfbadd58f3f400bc1b083a9158842"} Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.269982 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-utilities\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.270264 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-catalog-content\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.270289 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r42kx\" (UniqueName: \"kubernetes.io/projected/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-kube-api-access-r42kx\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.270510 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-utilities\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.270814 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-catalog-content\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.291278 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:32 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:32 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:32 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.291337 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.292311 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-grcdz"] Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.296736 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r42kx\" (UniqueName: \"kubernetes.io/projected/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-kube-api-access-r42kx\") pod \"redhat-operators-lvwbd\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: W1004 04:35:32.363164 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacebc5f9_1793_4b93_a1e9_55604bb69996.slice/crio-32e7d086e5a3f8ab09357a8d064a0565f24be01d6ba3587da26c926d8216ad06 WatchSource:0}: Error finding container 32e7d086e5a3f8ab09357a8d064a0565f24be01d6ba3587da26c926d8216ad06: Status 404 returned error can't find the container with id 32e7d086e5a3f8ab09357a8d064a0565f24be01d6ba3587da26c926d8216ad06 Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.419456 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:35:32 crc kubenswrapper[4992]: I1004 04:35:32.757249 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lvwbd"] Oct 04 04:35:32 crc kubenswrapper[4992]: W1004 04:35:32.763062 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7ad5d42_0b57_4e14_b972_51c0d2ac4270.slice/crio-c6b87d0995e4320a09dde6c275027befa48f5fdc6630a3df4ce86f9be9d5afd3 WatchSource:0}: Error finding container c6b87d0995e4320a09dde6c275027befa48f5fdc6630a3df4ce86f9be9d5afd3: Status 404 returned error can't find the container with id c6b87d0995e4320a09dde6c275027befa48f5fdc6630a3df4ce86f9be9d5afd3 Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.221153 4992 generic.go:334] "Generic (PLEG): container finished" podID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerID="fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2" exitCode=0 Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.221246 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpzhh" event={"ID":"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1","Type":"ContainerDied","Data":"fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2"} Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.224003 4992 generic.go:334] "Generic (PLEG): container finished" podID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerID="1105c2532fb9c9ea1da96965d86dd4dee35e07081bab24811c2896c3162f2b14" exitCode=0 Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.224080 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grcdz" event={"ID":"acebc5f9-1793-4b93-a1e9-55604bb69996","Type":"ContainerDied","Data":"1105c2532fb9c9ea1da96965d86dd4dee35e07081bab24811c2896c3162f2b14"} Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.224126 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grcdz" event={"ID":"acebc5f9-1793-4b93-a1e9-55604bb69996","Type":"ContainerStarted","Data":"32e7d086e5a3f8ab09357a8d064a0565f24be01d6ba3587da26c926d8216ad06"} Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.228057 4992 generic.go:334] "Generic (PLEG): container finished" podID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerID="de4b82b30a4b956401337f7f3cbaf1030fc0e67e9c020a0ff373a2db4f8e32a8" exitCode=0 Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.229317 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvwbd" event={"ID":"a7ad5d42-0b57-4e14-b972-51c0d2ac4270","Type":"ContainerDied","Data":"de4b82b30a4b956401337f7f3cbaf1030fc0e67e9c020a0ff373a2db4f8e32a8"} Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.229407 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvwbd" event={"ID":"a7ad5d42-0b57-4e14-b972-51c0d2ac4270","Type":"ContainerStarted","Data":"c6b87d0995e4320a09dde6c275027befa48f5fdc6630a3df4ce86f9be9d5afd3"} Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.287251 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:33 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:33 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:33 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.287297 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.511884 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.580791 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.689752 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0526e03e-5889-4202-a969-e2fec17fca8b-kube-api-access\") pod \"0526e03e-5889-4202-a969-e2fec17fca8b\" (UID: \"0526e03e-5889-4202-a969-e2fec17fca8b\") " Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.689848 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0526e03e-5889-4202-a969-e2fec17fca8b-kubelet-dir\") pod \"0526e03e-5889-4202-a969-e2fec17fca8b\" (UID: \"0526e03e-5889-4202-a969-e2fec17fca8b\") " Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.689908 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e72c23e9-d135-4282-801c-c6f4bddd318b-kube-api-access\") pod \"e72c23e9-d135-4282-801c-c6f4bddd318b\" (UID: \"e72c23e9-d135-4282-801c-c6f4bddd318b\") " Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.689974 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e72c23e9-d135-4282-801c-c6f4bddd318b-kubelet-dir\") pod \"e72c23e9-d135-4282-801c-c6f4bddd318b\" (UID: \"e72c23e9-d135-4282-801c-c6f4bddd318b\") " Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.690250 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e72c23e9-d135-4282-801c-c6f4bddd318b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e72c23e9-d135-4282-801c-c6f4bddd318b" (UID: "e72c23e9-d135-4282-801c-c6f4bddd318b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.690292 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0526e03e-5889-4202-a969-e2fec17fca8b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0526e03e-5889-4202-a969-e2fec17fca8b" (UID: "0526e03e-5889-4202-a969-e2fec17fca8b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.697081 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e72c23e9-d135-4282-801c-c6f4bddd318b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e72c23e9-d135-4282-801c-c6f4bddd318b" (UID: "e72c23e9-d135-4282-801c-c6f4bddd318b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.697193 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0526e03e-5889-4202-a969-e2fec17fca8b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0526e03e-5889-4202-a969-e2fec17fca8b" (UID: "0526e03e-5889-4202-a969-e2fec17fca8b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.791567 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0526e03e-5889-4202-a969-e2fec17fca8b-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.791607 4992 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0526e03e-5889-4202-a969-e2fec17fca8b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.791621 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e72c23e9-d135-4282-801c-c6f4bddd318b-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:33 crc kubenswrapper[4992]: I1004 04:35:33.791634 4992 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e72c23e9-d135-4282-801c-c6f4bddd318b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 04:35:34 crc kubenswrapper[4992]: I1004 04:35:34.250102 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0526e03e-5889-4202-a969-e2fec17fca8b","Type":"ContainerDied","Data":"d192b94227d2cac6e297703f64e1a1efbc181b7c676657566b0581df892e2d28"} Oct 04 04:35:34 crc kubenswrapper[4992]: I1004 04:35:34.250158 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d192b94227d2cac6e297703f64e1a1efbc181b7c676657566b0581df892e2d28" Oct 04 04:35:34 crc kubenswrapper[4992]: I1004 04:35:34.250230 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 04:35:34 crc kubenswrapper[4992]: I1004 04:35:34.259778 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e72c23e9-d135-4282-801c-c6f4bddd318b","Type":"ContainerDied","Data":"e2261788e2f7cff0b430cd3bc4ef3cec4e71baacbc2b5d529f3310881a6fa6ae"} Oct 04 04:35:34 crc kubenswrapper[4992]: I1004 04:35:34.259823 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2261788e2f7cff0b430cd3bc4ef3cec4e71baacbc2b5d529f3310881a6fa6ae" Oct 04 04:35:34 crc kubenswrapper[4992]: I1004 04:35:34.259883 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 04:35:34 crc kubenswrapper[4992]: I1004 04:35:34.286904 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:34 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:34 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:34 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:34 crc kubenswrapper[4992]: I1004 04:35:34.286966 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:35 crc kubenswrapper[4992]: I1004 04:35:35.288486 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:35 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:35 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:35 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:35 crc kubenswrapper[4992]: I1004 04:35:35.289122 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:35 crc kubenswrapper[4992]: I1004 04:35:35.345625 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:35:36 crc kubenswrapper[4992]: I1004 04:35:36.214879 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:36 crc kubenswrapper[4992]: I1004 04:35:36.221233 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pjtjn" Oct 04 04:35:36 crc kubenswrapper[4992]: I1004 04:35:36.291092 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:36 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:36 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:36 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:36 crc kubenswrapper[4992]: I1004 04:35:36.291151 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:36 crc kubenswrapper[4992]: I1004 04:35:36.725165 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-blzx5" Oct 04 04:35:37 crc kubenswrapper[4992]: I1004 04:35:37.287383 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:37 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:37 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:37 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:37 crc kubenswrapper[4992]: I1004 04:35:37.288101 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:37 crc kubenswrapper[4992]: I1004 04:35:37.770053 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:37 crc kubenswrapper[4992]: I1004 04:35:37.784132 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ef62768-7ed6-4ce3-9d34-6e6743681405-metrics-certs\") pod \"network-metrics-daemon-6jlms\" (UID: \"2ef62768-7ed6-4ce3-9d34-6e6743681405\") " pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:37 crc kubenswrapper[4992]: I1004 04:35:37.937437 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6jlms" Oct 04 04:35:38 crc kubenswrapper[4992]: I1004 04:35:38.286691 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:38 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:38 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:38 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:38 crc kubenswrapper[4992]: I1004 04:35:38.286772 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:39 crc kubenswrapper[4992]: I1004 04:35:39.287381 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:39 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:39 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:39 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:39 crc kubenswrapper[4992]: I1004 04:35:39.287466 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:40 crc kubenswrapper[4992]: I1004 04:35:40.286394 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:40 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:40 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:40 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:40 crc kubenswrapper[4992]: I1004 04:35:40.286455 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:41 crc kubenswrapper[4992]: I1004 04:35:41.173530 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:41 crc kubenswrapper[4992]: I1004 04:35:41.173593 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:41 crc kubenswrapper[4992]: I1004 04:35:41.173594 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:41 crc kubenswrapper[4992]: I1004 04:35:41.173689 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:41 crc kubenswrapper[4992]: I1004 04:35:41.176442 4992 patch_prober.go:28] interesting pod/console-f9d7485db-5sgbq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 04 04:35:41 crc kubenswrapper[4992]: I1004 04:35:41.176475 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5sgbq" podUID="8e339a50-fbf4-44da-9335-ba6ae033fc38" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 04 04:35:41 crc kubenswrapper[4992]: I1004 04:35:41.286493 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:41 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:41 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:41 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:41 crc kubenswrapper[4992]: I1004 04:35:41.286608 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:42 crc kubenswrapper[4992]: I1004 04:35:42.285816 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:42 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:42 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:42 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:42 crc kubenswrapper[4992]: I1004 04:35:42.285876 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:43 crc kubenswrapper[4992]: I1004 04:35:43.286672 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 04:35:43 crc kubenswrapper[4992]: [-]has-synced failed: reason withheld Oct 04 04:35:43 crc kubenswrapper[4992]: [+]process-running ok Oct 04 04:35:43 crc kubenswrapper[4992]: healthz check failed Oct 04 04:35:43 crc kubenswrapper[4992]: I1004 04:35:43.286873 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 04:35:44 crc kubenswrapper[4992]: I1004 04:35:44.286536 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:44 crc kubenswrapper[4992]: I1004 04:35:44.289994 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-68nfv" Oct 04 04:35:50 crc kubenswrapper[4992]: I1004 04:35:50.005017 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:35:50 crc kubenswrapper[4992]: I1004 04:35:50.894239 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:35:50 crc kubenswrapper[4992]: I1004 04:35:50.894564 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.173173 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.173230 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.173175 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.173628 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.173660 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.174269 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"1371c77d74650e0426a7b16f479a7a9595aa8c05934cc60da8142063f6b18f32"} pod="openshift-console/downloads-7954f5f757-r7t8c" containerMessage="Container download-server failed liveness probe, will be restarted" Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.174355 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" containerID="cri-o://1371c77d74650e0426a7b16f479a7a9595aa8c05934cc60da8142063f6b18f32" gracePeriod=2 Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.174795 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.174823 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.216025 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:51 crc kubenswrapper[4992]: I1004 04:35:51.222201 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:35:52 crc kubenswrapper[4992]: I1004 04:35:52.503707 4992 generic.go:334] "Generic (PLEG): container finished" podID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerID="1371c77d74650e0426a7b16f479a7a9595aa8c05934cc60da8142063f6b18f32" exitCode=0 Oct 04 04:35:52 crc kubenswrapper[4992]: I1004 04:35:52.503749 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r7t8c" event={"ID":"6a9ef22d-4511-403a-9703-fcffd8f5b8e9","Type":"ContainerDied","Data":"1371c77d74650e0426a7b16f479a7a9595aa8c05934cc60da8142063f6b18f32"} Oct 04 04:36:01 crc kubenswrapper[4992]: I1004 04:36:01.174314 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:01 crc kubenswrapper[4992]: I1004 04:36:01.176068 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:01 crc kubenswrapper[4992]: I1004 04:36:01.701302 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8c9n2" Oct 04 04:36:03 crc kubenswrapper[4992]: I1004 04:36:03.641911 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 04:36:11 crc kubenswrapper[4992]: I1004 04:36:11.173431 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:11 crc kubenswrapper[4992]: I1004 04:36:11.173876 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:20 crc kubenswrapper[4992]: I1004 04:36:20.894496 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:36:20 crc kubenswrapper[4992]: I1004 04:36:20.894976 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:36:20 crc kubenswrapper[4992]: I1004 04:36:20.895025 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:36:20 crc kubenswrapper[4992]: I1004 04:36:20.895579 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:36:20 crc kubenswrapper[4992]: I1004 04:36:20.895643 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef" gracePeriod=600 Oct 04 04:36:21 crc kubenswrapper[4992]: I1004 04:36:21.174074 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:21 crc kubenswrapper[4992]: I1004 04:36:21.174147 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:22 crc kubenswrapper[4992]: E1004 04:36:22.621396 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 04:36:22 crc kubenswrapper[4992]: E1004 04:36:22.621940 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r42kx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lvwbd_openshift-marketplace(a7ad5d42-0b57-4e14-b972-51c0d2ac4270): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:36:22 crc kubenswrapper[4992]: E1004 04:36:22.623204 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lvwbd" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" Oct 04 04:36:22 crc kubenswrapper[4992]: I1004 04:36:22.674562 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef" exitCode=0 Oct 04 04:36:22 crc kubenswrapper[4992]: I1004 04:36:22.674834 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef"} Oct 04 04:36:31 crc kubenswrapper[4992]: I1004 04:36:31.173569 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:31 crc kubenswrapper[4992]: I1004 04:36:31.174091 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:40 crc kubenswrapper[4992]: E1004 04:36:40.302338 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 04:36:40 crc kubenswrapper[4992]: E1004 04:36:40.303823 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4pg4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vwqln_openshift-marketplace(1463c32f-0eba-42da-b876-5c714cdf3c88): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:36:40 crc kubenswrapper[4992]: E1004 04:36:40.305113 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vwqln" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" Oct 04 04:36:41 crc kubenswrapper[4992]: I1004 04:36:41.173435 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:41 crc kubenswrapper[4992]: I1004 04:36:41.173495 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:41 crc kubenswrapper[4992]: E1004 04:36:41.412178 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 04:36:41 crc kubenswrapper[4992]: E1004 04:36:41.412437 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tfgx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-c4mz6_openshift-marketplace(cba17782-3829-4de8-b49e-9eba858fc1ed): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:36:41 crc kubenswrapper[4992]: E1004 04:36:41.413716 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-c4mz6" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" Oct 04 04:36:44 crc kubenswrapper[4992]: E1004 04:36:44.249291 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vwqln" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" Oct 04 04:36:44 crc kubenswrapper[4992]: E1004 04:36:44.249707 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-c4mz6" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" Oct 04 04:36:50 crc kubenswrapper[4992]: E1004 04:36:50.229683 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 04:36:50 crc kubenswrapper[4992]: E1004 04:36:50.230440 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-khdwl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-grcdz_openshift-marketplace(acebc5f9-1793-4b93-a1e9-55604bb69996): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:36:50 crc kubenswrapper[4992]: E1004 04:36:50.231579 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-grcdz" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" Oct 04 04:36:51 crc kubenswrapper[4992]: I1004 04:36:51.173070 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:36:51 crc kubenswrapper[4992]: I1004 04:36:51.173140 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:36:53 crc kubenswrapper[4992]: E1004 04:36:53.958326 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-grcdz" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" Oct 04 04:36:56 crc kubenswrapper[4992]: E1004 04:36:56.206499 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 04:36:56 crc kubenswrapper[4992]: E1004 04:36:56.206679 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nvhkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wp4ff_openshift-marketplace(5e7c9252-8288-416d-a5bb-0e1ffb586240): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:36:56 crc kubenswrapper[4992]: E1004 04:36:56.208010 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wp4ff" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" Oct 04 04:36:57 crc kubenswrapper[4992]: E1004 04:36:57.693404 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wp4ff" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" Oct 04 04:36:57 crc kubenswrapper[4992]: E1004 04:36:57.750308 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 04:36:57 crc kubenswrapper[4992]: E1004 04:36:57.750536 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2ld6g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-g8cmz_openshift-marketplace(35ec2da1-7bca-4056-96eb-67a9687c4472): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:36:57 crc kubenswrapper[4992]: E1004 04:36:57.751753 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-g8cmz" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" Oct 04 04:36:58 crc kubenswrapper[4992]: E1004 04:36:58.569630 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-g8cmz" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" Oct 04 04:36:58 crc kubenswrapper[4992]: I1004 04:36:58.755013 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6jlms"] Oct 04 04:36:58 crc kubenswrapper[4992]: E1004 04:36:58.821934 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 04:36:58 crc kubenswrapper[4992]: E1004 04:36:58.822086 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-82qwm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-w57r2_openshift-marketplace(151de3a2-278d-40d7-b869-023e9cd39273): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:36:58 crc kubenswrapper[4992]: E1004 04:36:58.823793 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-w57r2" podUID="151de3a2-278d-40d7-b869-023e9cd39273" Oct 04 04:36:59 crc kubenswrapper[4992]: E1004 04:36:59.237448 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 04:36:59 crc kubenswrapper[4992]: E1004 04:36:59.237647 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x8tpx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dpzhh_openshift-marketplace(1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 04:36:59 crc kubenswrapper[4992]: E1004 04:36:59.238869 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dpzhh" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" Oct 04 04:37:00 crc kubenswrapper[4992]: I1004 04:37:00.875929 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6jlms" event={"ID":"2ef62768-7ed6-4ce3-9d34-6e6743681405","Type":"ContainerStarted","Data":"e7748ef9b9fb4134fc56e75162047d1f94ca8dfa9125c21ef11ce2961bc4e694"} Oct 04 04:37:00 crc kubenswrapper[4992]: I1004 04:37:00.877870 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r7t8c" event={"ID":"6a9ef22d-4511-403a-9703-fcffd8f5b8e9","Type":"ContainerStarted","Data":"ce7110c9c556921f4fb3a7acafa27417ba3382ed8003b1b21c57a3b593f1329a"} Oct 04 04:37:00 crc kubenswrapper[4992]: I1004 04:37:00.878773 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:00 crc kubenswrapper[4992]: I1004 04:37:00.878817 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:00 crc kubenswrapper[4992]: I1004 04:37:00.878949 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:37:01 crc kubenswrapper[4992]: I1004 04:37:01.173179 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:01 crc kubenswrapper[4992]: I1004 04:37:01.173474 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:01 crc kubenswrapper[4992]: I1004 04:37:01.173234 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:01 crc kubenswrapper[4992]: I1004 04:37:01.173644 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:01 crc kubenswrapper[4992]: E1004 04:37:01.805960 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dpzhh" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" Oct 04 04:37:01 crc kubenswrapper[4992]: E1004 04:37:01.806290 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-w57r2" podUID="151de3a2-278d-40d7-b869-023e9cd39273" Oct 04 04:37:01 crc kubenswrapper[4992]: I1004 04:37:01.884966 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:01 crc kubenswrapper[4992]: I1004 04:37:01.885029 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:02 crc kubenswrapper[4992]: I1004 04:37:02.889818 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6jlms" event={"ID":"2ef62768-7ed6-4ce3-9d34-6e6743681405","Type":"ContainerStarted","Data":"f81cd3162e715cf9dce47515eda4073c2538ea6bf520897bc525cb3b2ed6cd53"} Oct 04 04:37:02 crc kubenswrapper[4992]: I1004 04:37:02.892660 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvwbd" event={"ID":"a7ad5d42-0b57-4e14-b972-51c0d2ac4270","Type":"ContainerStarted","Data":"ab151c7bd563d425a72283d0eba2c753dcf699d3cfae1de4dfbf04e77286fa74"} Oct 04 04:37:02 crc kubenswrapper[4992]: I1004 04:37:02.894665 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwqln" event={"ID":"1463c32f-0eba-42da-b876-5c714cdf3c88","Type":"ContainerStarted","Data":"4011ed42ad35952661f75f47c08e90ddc4be15da654e04576114af573deb47ef"} Oct 04 04:37:02 crc kubenswrapper[4992]: I1004 04:37:02.896443 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"93092393a842d35831bfc9be98a84a2f8a76369bad06835685afb28ae1178e48"} Oct 04 04:37:02 crc kubenswrapper[4992]: I1004 04:37:02.897121 4992 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7t8c container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Oct 04 04:37:02 crc kubenswrapper[4992]: I1004 04:37:02.897204 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7t8c" podUID="6a9ef22d-4511-403a-9703-fcffd8f5b8e9" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Oct 04 04:37:03 crc kubenswrapper[4992]: I1004 04:37:03.905990 4992 generic.go:334] "Generic (PLEG): container finished" podID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerID="4011ed42ad35952661f75f47c08e90ddc4be15da654e04576114af573deb47ef" exitCode=0 Oct 04 04:37:03 crc kubenswrapper[4992]: I1004 04:37:03.906117 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwqln" event={"ID":"1463c32f-0eba-42da-b876-5c714cdf3c88","Type":"ContainerDied","Data":"4011ed42ad35952661f75f47c08e90ddc4be15da654e04576114af573deb47ef"} Oct 04 04:37:03 crc kubenswrapper[4992]: I1004 04:37:03.910622 4992 generic.go:334] "Generic (PLEG): container finished" podID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerID="ab151c7bd563d425a72283d0eba2c753dcf699d3cfae1de4dfbf04e77286fa74" exitCode=0 Oct 04 04:37:03 crc kubenswrapper[4992]: I1004 04:37:03.910722 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvwbd" event={"ID":"a7ad5d42-0b57-4e14-b972-51c0d2ac4270","Type":"ContainerDied","Data":"ab151c7bd563d425a72283d0eba2c753dcf699d3cfae1de4dfbf04e77286fa74"} Oct 04 04:37:05 crc kubenswrapper[4992]: I1004 04:37:05.932301 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6jlms" event={"ID":"2ef62768-7ed6-4ce3-9d34-6e6743681405","Type":"ContainerStarted","Data":"160a9c84bbbc15f4338bc64ad64abd25e22b679d95377f2d63329d3cdb219405"} Oct 04 04:37:05 crc kubenswrapper[4992]: I1004 04:37:05.956072 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-6jlms" podStartSLOduration=230.956049354 podStartE2EDuration="3m50.956049354s" podCreationTimestamp="2025-10-04 04:33:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:37:05.955593001 +0000 UTC m=+259.803268479" watchObservedRunningTime="2025-10-04 04:37:05.956049354 +0000 UTC m=+259.803724852" Oct 04 04:37:10 crc kubenswrapper[4992]: I1004 04:37:10.955992 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mz6" event={"ID":"cba17782-3829-4de8-b49e-9eba858fc1ed","Type":"ContainerStarted","Data":"85c6ac5f60583a2db4bfe93391c94f84b8e9bbb74b11cc1d8a7bcac4803047e6"} Oct 04 04:37:11 crc kubenswrapper[4992]: I1004 04:37:11.194666 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-r7t8c" Oct 04 04:37:11 crc kubenswrapper[4992]: I1004 04:37:11.964218 4992 generic.go:334] "Generic (PLEG): container finished" podID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerID="85c6ac5f60583a2db4bfe93391c94f84b8e9bbb74b11cc1d8a7bcac4803047e6" exitCode=0 Oct 04 04:37:11 crc kubenswrapper[4992]: I1004 04:37:11.964309 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mz6" event={"ID":"cba17782-3829-4de8-b49e-9eba858fc1ed","Type":"ContainerDied","Data":"85c6ac5f60583a2db4bfe93391c94f84b8e9bbb74b11cc1d8a7bcac4803047e6"} Oct 04 04:37:12 crc kubenswrapper[4992]: I1004 04:37:12.325063 4992 patch_prober.go:28] interesting pod/router-default-5444994796-68nfv container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 04:37:12 crc kubenswrapper[4992]: I1004 04:37:12.325167 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-68nfv" podUID="80d7820e-ea64-433e-9842-9142a578f408" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:37:20 crc kubenswrapper[4992]: I1004 04:37:20.021201 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwqln" event={"ID":"1463c32f-0eba-42da-b876-5c714cdf3c88","Type":"ContainerStarted","Data":"b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463"} Oct 04 04:37:30 crc kubenswrapper[4992]: I1004 04:37:22.067879 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vwqln" podStartSLOduration=5.991777289 podStartE2EDuration="1m54.067847361s" podCreationTimestamp="2025-10-04 04:35:28 +0000 UTC" firstStartedPulling="2025-10-04 04:35:30.104217293 +0000 UTC m=+163.951892761" lastFinishedPulling="2025-10-04 04:37:18.180287365 +0000 UTC m=+272.027962833" observedRunningTime="2025-10-04 04:37:22.059563371 +0000 UTC m=+275.907238859" watchObservedRunningTime="2025-10-04 04:37:22.067847361 +0000 UTC m=+275.915522869" Oct 04 04:37:30 crc kubenswrapper[4992]: I1004 04:37:29.308578 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:37:30 crc kubenswrapper[4992]: I1004 04:37:29.309350 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:37:30 crc kubenswrapper[4992]: I1004 04:37:30.575584 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:37:30 crc kubenswrapper[4992]: I1004 04:37:30.654332 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:37:30 crc kubenswrapper[4992]: I1004 04:37:30.814165 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vwqln"] Oct 04 04:37:32 crc kubenswrapper[4992]: I1004 04:37:32.093390 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vwqln" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="registry-server" containerID="cri-o://b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463" gracePeriod=2 Oct 04 04:37:34 crc kubenswrapper[4992]: I1004 04:37:34.117687 4992 generic.go:334] "Generic (PLEG): container finished" podID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerID="b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463" exitCode=0 Oct 04 04:37:34 crc kubenswrapper[4992]: I1004 04:37:34.117926 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwqln" event={"ID":"1463c32f-0eba-42da-b876-5c714cdf3c88","Type":"ContainerDied","Data":"b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463"} Oct 04 04:37:39 crc kubenswrapper[4992]: E1004 04:37:39.308463 4992 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463 is running failed: container process not found" containerID="b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 04:37:39 crc kubenswrapper[4992]: E1004 04:37:39.309193 4992 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463 is running failed: container process not found" containerID="b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 04:37:39 crc kubenswrapper[4992]: E1004 04:37:39.309532 4992 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463 is running failed: container process not found" containerID="b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 04:37:39 crc kubenswrapper[4992]: E1004 04:37:39.309570 4992 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-vwqln" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="registry-server" Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.470499 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.576756 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pg4m\" (UniqueName: \"kubernetes.io/projected/1463c32f-0eba-42da-b876-5c714cdf3c88-kube-api-access-4pg4m\") pod \"1463c32f-0eba-42da-b876-5c714cdf3c88\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.576802 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-catalog-content\") pod \"1463c32f-0eba-42da-b876-5c714cdf3c88\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.576962 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-utilities\") pod \"1463c32f-0eba-42da-b876-5c714cdf3c88\" (UID: \"1463c32f-0eba-42da-b876-5c714cdf3c88\") " Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.577911 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-utilities" (OuterVolumeSpecName: "utilities") pod "1463c32f-0eba-42da-b876-5c714cdf3c88" (UID: "1463c32f-0eba-42da-b876-5c714cdf3c88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.595620 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1463c32f-0eba-42da-b876-5c714cdf3c88-kube-api-access-4pg4m" (OuterVolumeSpecName: "kube-api-access-4pg4m") pod "1463c32f-0eba-42da-b876-5c714cdf3c88" (UID: "1463c32f-0eba-42da-b876-5c714cdf3c88"). InnerVolumeSpecName "kube-api-access-4pg4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.628699 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1463c32f-0eba-42da-b876-5c714cdf3c88" (UID: "1463c32f-0eba-42da-b876-5c714cdf3c88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.678540 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.678574 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pg4m\" (UniqueName: \"kubernetes.io/projected/1463c32f-0eba-42da-b876-5c714cdf3c88-kube-api-access-4pg4m\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:39 crc kubenswrapper[4992]: I1004 04:37:39.678586 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1463c32f-0eba-42da-b876-5c714cdf3c88-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.153344 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w57r2" event={"ID":"151de3a2-278d-40d7-b869-023e9cd39273","Type":"ContainerStarted","Data":"062904d058ef5df016ad2a0c91e9341e5bd388133c89cc7b49c2e312a9ca45e2"} Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.155756 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mz6" event={"ID":"cba17782-3829-4de8-b49e-9eba858fc1ed","Type":"ContainerStarted","Data":"b28cb91e8efc5938ac8b74a3d74189d2a6611eaf51fb4485ab2368fc6636aaee"} Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.158790 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vwqln" event={"ID":"1463c32f-0eba-42da-b876-5c714cdf3c88","Type":"ContainerDied","Data":"4eca2325aa5397381a82ea7fc6b1e93e5a537b6a783420b792ba7919bcd6d449"} Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.158837 4992 scope.go:117] "RemoveContainer" containerID="b7efc0982385fdc5898e222f522bea8f003c31c9b0cb3bccd004c1197c92a463" Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.158988 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vwqln" Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.188291 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vwqln"] Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.192120 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vwqln"] Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.192177 4992 scope.go:117] "RemoveContainer" containerID="4011ed42ad35952661f75f47c08e90ddc4be15da654e04576114af573deb47ef" Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.224987 4992 scope.go:117] "RemoveContainer" containerID="81845d871d53bdbdbeacb241de7c3be39e959b7a1cc4b8d04c92b15b691bed9b" Oct 04 04:37:40 crc kubenswrapper[4992]: I1004 04:37:40.326062 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" path="/var/lib/kubelet/pods/1463c32f-0eba-42da-b876-5c714cdf3c88/volumes" Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.165225 4992 generic.go:334] "Generic (PLEG): container finished" podID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerID="82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435" exitCode=0 Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.165324 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpzhh" event={"ID":"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1","Type":"ContainerDied","Data":"82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435"} Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.170412 4992 generic.go:334] "Generic (PLEG): container finished" podID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerID="62c8fdf30cdaa5d3338620838bb44bf6ecd42f8d692096889670bb67e6f22be5" exitCode=0 Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.170472 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4ff" event={"ID":"5e7c9252-8288-416d-a5bb-0e1ffb586240","Type":"ContainerDied","Data":"62c8fdf30cdaa5d3338620838bb44bf6ecd42f8d692096889670bb67e6f22be5"} Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.173845 4992 generic.go:334] "Generic (PLEG): container finished" podID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerID="4b0f9c9fe1d8f451ce8fb31d569ed91ab3c2bfdc6413d51bf2c2d93d0ebf613b" exitCode=0 Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.173915 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8cmz" event={"ID":"35ec2da1-7bca-4056-96eb-67a9687c4472","Type":"ContainerDied","Data":"4b0f9c9fe1d8f451ce8fb31d569ed91ab3c2bfdc6413d51bf2c2d93d0ebf613b"} Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.177113 4992 generic.go:334] "Generic (PLEG): container finished" podID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerID="a725dbfaf05d3ec48ee35f147e381f25e00bd3c97c33d107f6b1a4c1308341c6" exitCode=0 Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.177197 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grcdz" event={"ID":"acebc5f9-1793-4b93-a1e9-55604bb69996","Type":"ContainerDied","Data":"a725dbfaf05d3ec48ee35f147e381f25e00bd3c97c33d107f6b1a4c1308341c6"} Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.181607 4992 generic.go:334] "Generic (PLEG): container finished" podID="151de3a2-278d-40d7-b869-023e9cd39273" containerID="062904d058ef5df016ad2a0c91e9341e5bd388133c89cc7b49c2e312a9ca45e2" exitCode=0 Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.181669 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w57r2" event={"ID":"151de3a2-278d-40d7-b869-023e9cd39273","Type":"ContainerDied","Data":"062904d058ef5df016ad2a0c91e9341e5bd388133c89cc7b49c2e312a9ca45e2"} Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.187200 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvwbd" event={"ID":"a7ad5d42-0b57-4e14-b972-51c0d2ac4270","Type":"ContainerStarted","Data":"2f5b95861b0df7eb6970eae6984b586a29f89e20f535efe49e27247ec45a90d0"} Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.239643 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c4mz6" podStartSLOduration=3.816565808 podStartE2EDuration="2m13.239624542s" podCreationTimestamp="2025-10-04 04:35:28 +0000 UTC" firstStartedPulling="2025-10-04 04:35:30.094387912 +0000 UTC m=+163.942063380" lastFinishedPulling="2025-10-04 04:37:39.517446646 +0000 UTC m=+293.365122114" observedRunningTime="2025-10-04 04:37:41.236115489 +0000 UTC m=+295.083790947" watchObservedRunningTime="2025-10-04 04:37:41.239624542 +0000 UTC m=+295.087300010" Oct 04 04:37:41 crc kubenswrapper[4992]: I1004 04:37:41.257957 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lvwbd" podStartSLOduration=2.820747006 podStartE2EDuration="2m9.257935109s" podCreationTimestamp="2025-10-04 04:35:32 +0000 UTC" firstStartedPulling="2025-10-04 04:35:33.231616525 +0000 UTC m=+167.079291993" lastFinishedPulling="2025-10-04 04:37:39.668804608 +0000 UTC m=+293.516480096" observedRunningTime="2025-10-04 04:37:41.25345753 +0000 UTC m=+295.101133008" watchObservedRunningTime="2025-10-04 04:37:41.257935109 +0000 UTC m=+295.105610577" Oct 04 04:37:42 crc kubenswrapper[4992]: I1004 04:37:42.419930 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:37:42 crc kubenswrapper[4992]: I1004 04:37:42.420300 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:37:43 crc kubenswrapper[4992]: I1004 04:37:43.460804 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lvwbd" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="registry-server" probeResult="failure" output=< Oct 04 04:37:43 crc kubenswrapper[4992]: timeout: failed to connect service ":50051" within 1s Oct 04 04:37:43 crc kubenswrapper[4992]: > Oct 04 04:37:48 crc kubenswrapper[4992]: I1004 04:37:48.850786 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:37:48 crc kubenswrapper[4992]: I1004 04:37:48.851353 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:37:48 crc kubenswrapper[4992]: I1004 04:37:48.893058 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:37:49 crc kubenswrapper[4992]: I1004 04:37:49.275740 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:37:52 crc kubenswrapper[4992]: I1004 04:37:52.255796 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4ff" event={"ID":"5e7c9252-8288-416d-a5bb-0e1ffb586240","Type":"ContainerStarted","Data":"1e2b23304c018a244064df5a414daed70ebc2ea6dad5686894d736eabc3efd6b"} Oct 04 04:37:52 crc kubenswrapper[4992]: I1004 04:37:52.496473 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:37:52 crc kubenswrapper[4992]: I1004 04:37:52.575901 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:37:52 crc kubenswrapper[4992]: I1004 04:37:52.737129 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lvwbd"] Oct 04 04:37:53 crc kubenswrapper[4992]: I1004 04:37:53.279611 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wp4ff" podStartSLOduration=5.9113977030000004 podStartE2EDuration="2m24.279592169s" podCreationTimestamp="2025-10-04 04:35:29 +0000 UTC" firstStartedPulling="2025-10-04 04:35:31.179903702 +0000 UTC m=+165.027579170" lastFinishedPulling="2025-10-04 04:37:49.548098158 +0000 UTC m=+303.395773636" observedRunningTime="2025-10-04 04:37:53.275205063 +0000 UTC m=+307.122880531" watchObservedRunningTime="2025-10-04 04:37:53.279592169 +0000 UTC m=+307.127267637" Oct 04 04:37:54 crc kubenswrapper[4992]: I1004 04:37:54.265568 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lvwbd" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="registry-server" containerID="cri-o://2f5b95861b0df7eb6970eae6984b586a29f89e20f535efe49e27247ec45a90d0" gracePeriod=2 Oct 04 04:37:57 crc kubenswrapper[4992]: I1004 04:37:57.283475 4992 generic.go:334] "Generic (PLEG): container finished" podID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerID="2f5b95861b0df7eb6970eae6984b586a29f89e20f535efe49e27247ec45a90d0" exitCode=0 Oct 04 04:37:57 crc kubenswrapper[4992]: I1004 04:37:57.283557 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvwbd" event={"ID":"a7ad5d42-0b57-4e14-b972-51c0d2ac4270","Type":"ContainerDied","Data":"2f5b95861b0df7eb6970eae6984b586a29f89e20f535efe49e27247ec45a90d0"} Oct 04 04:37:59 crc kubenswrapper[4992]: I1004 04:37:59.481218 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:37:59 crc kubenswrapper[4992]: I1004 04:37:59.481665 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:37:59 crc kubenswrapper[4992]: I1004 04:37:59.520411 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:38:00 crc kubenswrapper[4992]: I1004 04:38:00.341464 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:38:00 crc kubenswrapper[4992]: I1004 04:38:00.383516 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wp4ff"] Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.577435 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.745040 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-catalog-content\") pod \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.745216 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-utilities\") pod \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.745331 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r42kx\" (UniqueName: \"kubernetes.io/projected/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-kube-api-access-r42kx\") pod \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\" (UID: \"a7ad5d42-0b57-4e14-b972-51c0d2ac4270\") " Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.746129 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-utilities" (OuterVolumeSpecName: "utilities") pod "a7ad5d42-0b57-4e14-b972-51c0d2ac4270" (UID: "a7ad5d42-0b57-4e14-b972-51c0d2ac4270"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.750707 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-kube-api-access-r42kx" (OuterVolumeSpecName: "kube-api-access-r42kx") pod "a7ad5d42-0b57-4e14-b972-51c0d2ac4270" (UID: "a7ad5d42-0b57-4e14-b972-51c0d2ac4270"). InnerVolumeSpecName "kube-api-access-r42kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.846581 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.846632 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r42kx\" (UniqueName: \"kubernetes.io/projected/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-kube-api-access-r42kx\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.861382 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7ad5d42-0b57-4e14-b972-51c0d2ac4270" (UID: "a7ad5d42-0b57-4e14-b972-51c0d2ac4270"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:38:01 crc kubenswrapper[4992]: I1004 04:38:01.948297 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7ad5d42-0b57-4e14-b972-51c0d2ac4270-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:02 crc kubenswrapper[4992]: I1004 04:38:02.311902 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lvwbd" event={"ID":"a7ad5d42-0b57-4e14-b972-51c0d2ac4270","Type":"ContainerDied","Data":"c6b87d0995e4320a09dde6c275027befa48f5fdc6630a3df4ce86f9be9d5afd3"} Oct 04 04:38:02 crc kubenswrapper[4992]: I1004 04:38:02.311953 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lvwbd" Oct 04 04:38:02 crc kubenswrapper[4992]: I1004 04:38:02.311983 4992 scope.go:117] "RemoveContainer" containerID="2f5b95861b0df7eb6970eae6984b586a29f89e20f535efe49e27247ec45a90d0" Oct 04 04:38:02 crc kubenswrapper[4992]: I1004 04:38:02.312037 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wp4ff" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerName="registry-server" containerID="cri-o://1e2b23304c018a244064df5a414daed70ebc2ea6dad5686894d736eabc3efd6b" gracePeriod=2 Oct 04 04:38:02 crc kubenswrapper[4992]: I1004 04:38:02.354893 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lvwbd"] Oct 04 04:38:02 crc kubenswrapper[4992]: I1004 04:38:02.359869 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lvwbd"] Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.104215 4992 scope.go:117] "RemoveContainer" containerID="ab151c7bd563d425a72283d0eba2c753dcf699d3cfae1de4dfbf04e77286fa74" Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.149349 4992 scope.go:117] "RemoveContainer" containerID="de4b82b30a4b956401337f7f3cbaf1030fc0e67e9c020a0ff373a2db4f8e32a8" Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.320206 4992 generic.go:334] "Generic (PLEG): container finished" podID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerID="1e2b23304c018a244064df5a414daed70ebc2ea6dad5686894d736eabc3efd6b" exitCode=0 Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.320284 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4ff" event={"ID":"5e7c9252-8288-416d-a5bb-0e1ffb586240","Type":"ContainerDied","Data":"1e2b23304c018a244064df5a414daed70ebc2ea6dad5686894d736eabc3efd6b"} Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.536792 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.671088 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-utilities\") pod \"5e7c9252-8288-416d-a5bb-0e1ffb586240\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.671164 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvhkk\" (UniqueName: \"kubernetes.io/projected/5e7c9252-8288-416d-a5bb-0e1ffb586240-kube-api-access-nvhkk\") pod \"5e7c9252-8288-416d-a5bb-0e1ffb586240\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.671220 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-catalog-content\") pod \"5e7c9252-8288-416d-a5bb-0e1ffb586240\" (UID: \"5e7c9252-8288-416d-a5bb-0e1ffb586240\") " Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.672399 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-utilities" (OuterVolumeSpecName: "utilities") pod "5e7c9252-8288-416d-a5bb-0e1ffb586240" (UID: "5e7c9252-8288-416d-a5bb-0e1ffb586240"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.683658 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e7c9252-8288-416d-a5bb-0e1ffb586240-kube-api-access-nvhkk" (OuterVolumeSpecName: "kube-api-access-nvhkk") pod "5e7c9252-8288-416d-a5bb-0e1ffb586240" (UID: "5e7c9252-8288-416d-a5bb-0e1ffb586240"). InnerVolumeSpecName "kube-api-access-nvhkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.728983 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e7c9252-8288-416d-a5bb-0e1ffb586240" (UID: "5e7c9252-8288-416d-a5bb-0e1ffb586240"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.772710 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.772765 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvhkk\" (UniqueName: \"kubernetes.io/projected/5e7c9252-8288-416d-a5bb-0e1ffb586240-kube-api-access-nvhkk\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:03 crc kubenswrapper[4992]: I1004 04:38:03.772777 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e7c9252-8288-416d-a5bb-0e1ffb586240-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.326508 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" path="/var/lib/kubelet/pods/a7ad5d42-0b57-4e14-b972-51c0d2ac4270/volumes" Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.327646 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w57r2" event={"ID":"151de3a2-278d-40d7-b869-023e9cd39273","Type":"ContainerStarted","Data":"7b2b6ab6417242dcf4b97a03d8a7a76f7d59029aa8d16535d2fbbac74da228b0"} Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.334222 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpzhh" event={"ID":"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1","Type":"ContainerStarted","Data":"c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961"} Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.336266 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4ff" event={"ID":"5e7c9252-8288-416d-a5bb-0e1ffb586240","Type":"ContainerDied","Data":"22e690f0338c1aeb3a8ef2408a67987e33b835d97f9534de85a9922406b53389"} Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.336297 4992 scope.go:117] "RemoveContainer" containerID="1e2b23304c018a244064df5a414daed70ebc2ea6dad5686894d736eabc3efd6b" Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.336407 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp4ff" Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.341394 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8cmz" event={"ID":"35ec2da1-7bca-4056-96eb-67a9687c4472","Type":"ContainerStarted","Data":"4053d7680d4a07f906b4cca19390824bbaa0af958fd22db3253dd80a46bb1868"} Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.346917 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grcdz" event={"ID":"acebc5f9-1793-4b93-a1e9-55604bb69996","Type":"ContainerStarted","Data":"6ff7a09c2a3bccd4fbab24f05a4e6c0ec632dd893bd7ff90579f08d8f1e6803d"} Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.369300 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g8cmz" podStartSLOduration=3.334631634 podStartE2EDuration="2m36.369286466s" podCreationTimestamp="2025-10-04 04:35:28 +0000 UTC" firstStartedPulling="2025-10-04 04:35:30.069534202 +0000 UTC m=+163.917209670" lastFinishedPulling="2025-10-04 04:38:03.104189034 +0000 UTC m=+316.951864502" observedRunningTime="2025-10-04 04:38:04.369112412 +0000 UTC m=+318.216787880" watchObservedRunningTime="2025-10-04 04:38:04.369286466 +0000 UTC m=+318.216961934" Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.369440 4992 scope.go:117] "RemoveContainer" containerID="62c8fdf30cdaa5d3338620838bb44bf6ecd42f8d692096889670bb67e6f22be5" Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.384928 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wp4ff"] Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.387965 4992 scope.go:117] "RemoveContainer" containerID="8fb474035d2fe0efc33beb530f362487049611706ae632e06080f51e7d8e8824" Oct 04 04:38:04 crc kubenswrapper[4992]: I1004 04:38:04.388014 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wp4ff"] Oct 04 04:38:05 crc kubenswrapper[4992]: I1004 04:38:05.377253 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w57r2" podStartSLOduration=4.484458051 podStartE2EDuration="2m35.377181954s" podCreationTimestamp="2025-10-04 04:35:30 +0000 UTC" firstStartedPulling="2025-10-04 04:35:32.217962643 +0000 UTC m=+166.065638111" lastFinishedPulling="2025-10-04 04:38:03.110686546 +0000 UTC m=+316.958362014" observedRunningTime="2025-10-04 04:38:05.375758366 +0000 UTC m=+319.223433894" watchObservedRunningTime="2025-10-04 04:38:05.377181954 +0000 UTC m=+319.224857462" Oct 04 04:38:05 crc kubenswrapper[4992]: I1004 04:38:05.407591 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-grcdz" podStartSLOduration=4.525685214 podStartE2EDuration="2m34.407551161s" podCreationTimestamp="2025-10-04 04:35:31 +0000 UTC" firstStartedPulling="2025-10-04 04:35:33.22618966 +0000 UTC m=+167.073865128" lastFinishedPulling="2025-10-04 04:38:03.108055597 +0000 UTC m=+316.955731075" observedRunningTime="2025-10-04 04:38:05.401254334 +0000 UTC m=+319.248929802" watchObservedRunningTime="2025-10-04 04:38:05.407551161 +0000 UTC m=+319.255226719" Oct 04 04:38:05 crc kubenswrapper[4992]: I1004 04:38:05.431746 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dpzhh" podStartSLOduration=5.102731436 podStartE2EDuration="2m34.431720373s" podCreationTimestamp="2025-10-04 04:35:31 +0000 UTC" firstStartedPulling="2025-10-04 04:35:32.189572538 +0000 UTC m=+166.037248006" lastFinishedPulling="2025-10-04 04:38:01.518561465 +0000 UTC m=+315.366236943" observedRunningTime="2025-10-04 04:38:05.428679882 +0000 UTC m=+319.276355360" watchObservedRunningTime="2025-10-04 04:38:05.431720373 +0000 UTC m=+319.279395851" Oct 04 04:38:06 crc kubenswrapper[4992]: I1004 04:38:06.334751 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" path="/var/lib/kubelet/pods/5e7c9252-8288-416d-a5bb-0e1ffb586240/volumes" Oct 04 04:38:08 crc kubenswrapper[4992]: I1004 04:38:08.994661 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:38:08 crc kubenswrapper[4992]: I1004 04:38:08.995298 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:38:09 crc kubenswrapper[4992]: I1004 04:38:09.057476 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:38:09 crc kubenswrapper[4992]: I1004 04:38:09.419945 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:38:10 crc kubenswrapper[4992]: I1004 04:38:10.973047 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:38:10 crc kubenswrapper[4992]: I1004 04:38:10.973597 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:38:11 crc kubenswrapper[4992]: I1004 04:38:11.014426 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:38:11 crc kubenswrapper[4992]: I1004 04:38:11.388977 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:38:11 crc kubenswrapper[4992]: I1004 04:38:11.389224 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:38:11 crc kubenswrapper[4992]: I1004 04:38:11.426460 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:38:11 crc kubenswrapper[4992]: I1004 04:38:11.433042 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:38:12 crc kubenswrapper[4992]: I1004 04:38:12.022235 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:38:12 crc kubenswrapper[4992]: I1004 04:38:12.022300 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:38:12 crc kubenswrapper[4992]: I1004 04:38:12.066391 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:38:12 crc kubenswrapper[4992]: I1004 04:38:12.441730 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:38:12 crc kubenswrapper[4992]: I1004 04:38:12.454069 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:38:13 crc kubenswrapper[4992]: I1004 04:38:13.749397 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpzhh"] Oct 04 04:38:15 crc kubenswrapper[4992]: I1004 04:38:15.412258 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dpzhh" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerName="registry-server" containerID="cri-o://c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961" gracePeriod=2 Oct 04 04:38:15 crc kubenswrapper[4992]: I1004 04:38:15.812798 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:38:15 crc kubenswrapper[4992]: I1004 04:38:15.929914 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-catalog-content\") pod \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " Oct 04 04:38:15 crc kubenswrapper[4992]: I1004 04:38:15.930723 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-utilities\") pod \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " Oct 04 04:38:15 crc kubenswrapper[4992]: I1004 04:38:15.930803 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8tpx\" (UniqueName: \"kubernetes.io/projected/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-kube-api-access-x8tpx\") pod \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\" (UID: \"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1\") " Oct 04 04:38:15 crc kubenswrapper[4992]: I1004 04:38:15.931539 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-utilities" (OuterVolumeSpecName: "utilities") pod "1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" (UID: "1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:38:15 crc kubenswrapper[4992]: I1004 04:38:15.936292 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-kube-api-access-x8tpx" (OuterVolumeSpecName: "kube-api-access-x8tpx") pod "1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" (UID: "1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1"). InnerVolumeSpecName "kube-api-access-x8tpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:38:15 crc kubenswrapper[4992]: I1004 04:38:15.942223 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" (UID: "1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.031764 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8tpx\" (UniqueName: \"kubernetes.io/projected/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-kube-api-access-x8tpx\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.031808 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.031818 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.418991 4992 generic.go:334] "Generic (PLEG): container finished" podID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerID="c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961" exitCode=0 Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.419032 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpzhh" event={"ID":"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1","Type":"ContainerDied","Data":"c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961"} Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.419062 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dpzhh" event={"ID":"1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1","Type":"ContainerDied","Data":"15bd86efac51b7329ad1325fdf10f192841b09b107e45bb70543645d244126ca"} Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.419077 4992 scope.go:117] "RemoveContainer" containerID="c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.419082 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dpzhh" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.435659 4992 scope.go:117] "RemoveContainer" containerID="82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.437490 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpzhh"] Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.440124 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dpzhh"] Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.452223 4992 scope.go:117] "RemoveContainer" containerID="fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.464579 4992 scope.go:117] "RemoveContainer" containerID="c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961" Oct 04 04:38:16 crc kubenswrapper[4992]: E1004 04:38:16.465165 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961\": container with ID starting with c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961 not found: ID does not exist" containerID="c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.465201 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961"} err="failed to get container status \"c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961\": rpc error: code = NotFound desc = could not find container \"c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961\": container with ID starting with c902f71cf003b7235166b4b17fef349c5f1169f974e1fddd6add2d545df5a961 not found: ID does not exist" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.465234 4992 scope.go:117] "RemoveContainer" containerID="82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435" Oct 04 04:38:16 crc kubenswrapper[4992]: E1004 04:38:16.465706 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435\": container with ID starting with 82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435 not found: ID does not exist" containerID="82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.465756 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435"} err="failed to get container status \"82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435\": rpc error: code = NotFound desc = could not find container \"82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435\": container with ID starting with 82f970a8fc9a5ab47b28ecddebfa7a59162c25cfe6beca73db9f421c5f747435 not found: ID does not exist" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.465788 4992 scope.go:117] "RemoveContainer" containerID="fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2" Oct 04 04:38:16 crc kubenswrapper[4992]: E1004 04:38:16.466131 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2\": container with ID starting with fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2 not found: ID does not exist" containerID="fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2" Oct 04 04:38:16 crc kubenswrapper[4992]: I1004 04:38:16.466159 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2"} err="failed to get container status \"fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2\": rpc error: code = NotFound desc = could not find container \"fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2\": container with ID starting with fc1d85ec8ad409482d632557f7c7c50a52d5a17bed4e52db16f080ff89dce2c2 not found: ID does not exist" Oct 04 04:38:18 crc kubenswrapper[4992]: I1004 04:38:18.325409 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" path="/var/lib/kubelet/pods/1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1/volumes" Oct 04 04:38:52 crc kubenswrapper[4992]: I1004 04:38:52.148727 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rd62r"] Oct 04 04:39:17 crc kubenswrapper[4992]: I1004 04:39:17.188131 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" podUID="dd501849-1d14-48c1-bc51-06d23c56db28" containerName="oauth-openshift" containerID="cri-o://c0e948ba0c4d699e3726f94988295b17571d1279775fbe46aa70fe01836a16f7" gracePeriod=15 Oct 04 04:39:17 crc kubenswrapper[4992]: I1004 04:39:17.797773 4992 generic.go:334] "Generic (PLEG): container finished" podID="dd501849-1d14-48c1-bc51-06d23c56db28" containerID="c0e948ba0c4d699e3726f94988295b17571d1279775fbe46aa70fe01836a16f7" exitCode=0 Oct 04 04:39:17 crc kubenswrapper[4992]: I1004 04:39:17.797825 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" event={"ID":"dd501849-1d14-48c1-bc51-06d23c56db28","Type":"ContainerDied","Data":"c0e948ba0c4d699e3726f94988295b17571d1279775fbe46aa70fe01836a16f7"} Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.047444 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080220 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt"] Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080442 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerName="extract-utilities" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080454 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerName="extract-utilities" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080463 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e72c23e9-d135-4282-801c-c6f4bddd318b" containerName="pruner" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080469 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="e72c23e9-d135-4282-801c-c6f4bddd318b" containerName="pruner" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080478 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080484 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080491 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080496 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080504 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0526e03e-5889-4202-a969-e2fec17fca8b" containerName="pruner" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080510 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="0526e03e-5889-4202-a969-e2fec17fca8b" containerName="pruner" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080519 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="extract-utilities" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080525 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="extract-utilities" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080531 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="extract-content" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080537 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="extract-content" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080543 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080550 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080557 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerName="extract-content" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080563 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerName="extract-content" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080574 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd501849-1d14-48c1-bc51-06d23c56db28" containerName="oauth-openshift" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080580 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd501849-1d14-48c1-bc51-06d23c56db28" containerName="oauth-openshift" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080588 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="extract-utilities" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080593 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="extract-utilities" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080600 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerName="extract-content" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080605 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerName="extract-content" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080617 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerName="extract-utilities" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080622 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerName="extract-utilities" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080632 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080638 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: E1004 04:39:18.080644 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="extract-content" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080650 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="extract-content" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080739 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd501849-1d14-48c1-bc51-06d23c56db28" containerName="oauth-openshift" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080749 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="e72c23e9-d135-4282-801c-c6f4bddd318b" containerName="pruner" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080757 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="1463c32f-0eba-42da-b876-5c714cdf3c88" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080766 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="0526e03e-5889-4202-a969-e2fec17fca8b" containerName="pruner" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080779 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea2c4fc-4545-4f6d-85f9-20a8d9ca87f1" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080787 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7ad5d42-0b57-4e14-b972-51c0d2ac4270" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.080796 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7c9252-8288-416d-a5bb-0e1ffb586240" containerName="registry-server" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.081182 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.108672 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt"] Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130323 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-audit-policies\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130434 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-router-certs\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130455 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-cliconfig\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130473 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-provider-selection\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130489 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-session\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130508 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-serving-cert\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130534 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-error\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130556 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-idp-0-file-data\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130576 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-login\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130593 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-service-ca\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130613 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-ocp-branding-template\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130632 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-trusted-ca-bundle\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130660 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hmmk\" (UniqueName: \"kubernetes.io/projected/dd501849-1d14-48c1-bc51-06d23c56db28-kube-api-access-9hmmk\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130679 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd501849-1d14-48c1-bc51-06d23c56db28-audit-dir\") pod \"dd501849-1d14-48c1-bc51-06d23c56db28\" (UID: \"dd501849-1d14-48c1-bc51-06d23c56db28\") " Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130809 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgqw4\" (UniqueName: \"kubernetes.io/projected/4db24f62-8b38-4b61-bdd5-76c38520f0d5-kube-api-access-mgqw4\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130833 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130849 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-service-ca\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130866 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4db24f62-8b38-4b61-bdd5-76c38520f0d5-audit-dir\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130892 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-session\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130906 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130923 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130939 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-router-certs\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130966 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.130986 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.131000 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-error\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.131000 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.131021 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-login\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.131040 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.131052 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.131058 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-audit-policies\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.131210 4992 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.131228 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.134108 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dd501849-1d14-48c1-bc51-06d23c56db28-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.134590 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.135931 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.136183 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.137549 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.138318 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd501849-1d14-48c1-bc51-06d23c56db28-kube-api-access-9hmmk" (OuterVolumeSpecName: "kube-api-access-9hmmk") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "kube-api-access-9hmmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.138581 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.153562 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.154007 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.154301 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.154395 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.165408 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "dd501849-1d14-48c1-bc51-06d23c56db28" (UID: "dd501849-1d14-48c1-bc51-06d23c56db28"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.231819 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-audit-policies\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.231891 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgqw4\" (UniqueName: \"kubernetes.io/projected/4db24f62-8b38-4b61-bdd5-76c38520f0d5-kube-api-access-mgqw4\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.231912 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.231928 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-service-ca\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.231947 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4db24f62-8b38-4b61-bdd5-76c38520f0d5-audit-dir\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.231972 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-session\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.231988 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232006 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-router-certs\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232053 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232094 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232111 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232106 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4db24f62-8b38-4b61-bdd5-76c38520f0d5-audit-dir\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232129 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-error\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232285 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-login\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232323 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232436 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232452 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232466 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232480 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232495 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232509 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232524 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232537 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hmmk\" (UniqueName: \"kubernetes.io/projected/dd501849-1d14-48c1-bc51-06d23c56db28-kube-api-access-9hmmk\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232550 4992 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dd501849-1d14-48c1-bc51-06d23c56db28-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232568 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232583 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232598 4992 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dd501849-1d14-48c1-bc51-06d23c56db28-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232853 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-audit-policies\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.232853 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-service-ca\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.233680 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.234181 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.235525 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.235551 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-session\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.235559 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-login\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.235915 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.236170 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.236333 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-user-template-error\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.236787 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.237332 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4db24f62-8b38-4b61-bdd5-76c38520f0d5-v4-0-config-system-router-certs\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.251742 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgqw4\" (UniqueName: \"kubernetes.io/projected/4db24f62-8b38-4b61-bdd5-76c38520f0d5-kube-api-access-mgqw4\") pod \"oauth-openshift-58b6dc46cc-ml5qt\" (UID: \"4db24f62-8b38-4b61-bdd5-76c38520f0d5\") " pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.408415 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.804701 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" event={"ID":"dd501849-1d14-48c1-bc51-06d23c56db28","Type":"ContainerDied","Data":"895c15af4ca9601edb7369b96e0076e104933d1e855eab74dbb7799280877b72"} Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.805146 4992 scope.go:117] "RemoveContainer" containerID="c0e948ba0c4d699e3726f94988295b17571d1279775fbe46aa70fe01836a16f7" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.804778 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-rd62r" Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.825947 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rd62r"] Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.829717 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-rd62r"] Oct 04 04:39:18 crc kubenswrapper[4992]: I1004 04:39:18.852029 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt"] Oct 04 04:39:18 crc kubenswrapper[4992]: W1004 04:39:18.859163 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4db24f62_8b38_4b61_bdd5_76c38520f0d5.slice/crio-2ecccde4ecfaf93c17aae4bb8b639aa3c0a9a28eddd00b6c99c2d087dabc7c3c WatchSource:0}: Error finding container 2ecccde4ecfaf93c17aae4bb8b639aa3c0a9a28eddd00b6c99c2d087dabc7c3c: Status 404 returned error can't find the container with id 2ecccde4ecfaf93c17aae4bb8b639aa3c0a9a28eddd00b6c99c2d087dabc7c3c Oct 04 04:39:19 crc kubenswrapper[4992]: I1004 04:39:19.813316 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" event={"ID":"4db24f62-8b38-4b61-bdd5-76c38520f0d5","Type":"ContainerStarted","Data":"e4a5c0f570f188fb8499797d4548966fc154e1f2cb8f6cf4a1815b000a73b39e"} Oct 04 04:39:19 crc kubenswrapper[4992]: I1004 04:39:19.813622 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:19 crc kubenswrapper[4992]: I1004 04:39:19.813725 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" event={"ID":"4db24f62-8b38-4b61-bdd5-76c38520f0d5","Type":"ContainerStarted","Data":"2ecccde4ecfaf93c17aae4bb8b639aa3c0a9a28eddd00b6c99c2d087dabc7c3c"} Oct 04 04:39:19 crc kubenswrapper[4992]: I1004 04:39:19.818831 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" Oct 04 04:39:19 crc kubenswrapper[4992]: I1004 04:39:19.833484 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-58b6dc46cc-ml5qt" podStartSLOduration=27.833460305 podStartE2EDuration="27.833460305s" podCreationTimestamp="2025-10-04 04:38:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:39:19.832066417 +0000 UTC m=+393.679741895" watchObservedRunningTime="2025-10-04 04:39:19.833460305 +0000 UTC m=+393.681135773" Oct 04 04:39:20 crc kubenswrapper[4992]: I1004 04:39:20.326266 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd501849-1d14-48c1-bc51-06d23c56db28" path="/var/lib/kubelet/pods/dd501849-1d14-48c1-bc51-06d23c56db28/volumes" Oct 04 04:39:20 crc kubenswrapper[4992]: I1004 04:39:20.894718 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:39:20 crc kubenswrapper[4992]: I1004 04:39:20.895271 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.137112 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4mz6"] Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.139308 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c4mz6" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerName="registry-server" containerID="cri-o://b28cb91e8efc5938ac8b74a3d74189d2a6611eaf51fb4485ab2368fc6636aaee" gracePeriod=30 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.140414 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g8cmz"] Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.140623 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g8cmz" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerName="registry-server" containerID="cri-o://4053d7680d4a07f906b4cca19390824bbaa0af958fd22db3253dd80a46bb1868" gracePeriod=30 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.162163 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tz4gp"] Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.162416 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" podUID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" containerName="marketplace-operator" containerID="cri-o://27ead3d3f43bdede09c3827abb6f736a3996fe21b3dbc48dfa59cd63e6e2d29a" gracePeriod=30 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.171993 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w57r2"] Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.172376 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w57r2" podUID="151de3a2-278d-40d7-b869-023e9cd39273" containerName="registry-server" containerID="cri-o://7b2b6ab6417242dcf4b97a03d8a7a76f7d59029aa8d16535d2fbbac74da228b0" gracePeriod=30 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.173533 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-grcdz"] Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.173872 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-grcdz" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerName="registry-server" containerID="cri-o://6ff7a09c2a3bccd4fbab24f05a4e6c0ec632dd893bd7ff90579f08d8f1e6803d" gracePeriod=30 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.181397 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ztwc8"] Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.182198 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.183901 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ztwc8"] Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.278572 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8584f53-da6d-4606-8ac5-557d0fa24828-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.278620 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78cfr\" (UniqueName: \"kubernetes.io/projected/c8584f53-da6d-4606-8ac5-557d0fa24828-kube-api-access-78cfr\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.278668 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8584f53-da6d-4606-8ac5-557d0fa24828-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.379631 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8584f53-da6d-4606-8ac5-557d0fa24828-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.379771 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8584f53-da6d-4606-8ac5-557d0fa24828-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.379798 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78cfr\" (UniqueName: \"kubernetes.io/projected/c8584f53-da6d-4606-8ac5-557d0fa24828-kube-api-access-78cfr\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.382848 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c8584f53-da6d-4606-8ac5-557d0fa24828-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.386074 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c8584f53-da6d-4606-8ac5-557d0fa24828-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.398059 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78cfr\" (UniqueName: \"kubernetes.io/projected/c8584f53-da6d-4606-8ac5-557d0fa24828-kube-api-access-78cfr\") pod \"marketplace-operator-79b997595-ztwc8\" (UID: \"c8584f53-da6d-4606-8ac5-557d0fa24828\") " pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.541783 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.746138 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ztwc8"] Oct 04 04:39:33 crc kubenswrapper[4992]: W1004 04:39:33.831490 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8584f53_da6d_4606_8ac5_557d0fa24828.slice/crio-df83ddca6417d2381aac48839703e6accab0085bfcb7d1403d5ca29b61f632d6 WatchSource:0}: Error finding container df83ddca6417d2381aac48839703e6accab0085bfcb7d1403d5ca29b61f632d6: Status 404 returned error can't find the container with id df83ddca6417d2381aac48839703e6accab0085bfcb7d1403d5ca29b61f632d6 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.894540 4992 generic.go:334] "Generic (PLEG): container finished" podID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerID="b28cb91e8efc5938ac8b74a3d74189d2a6611eaf51fb4485ab2368fc6636aaee" exitCode=0 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.894582 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mz6" event={"ID":"cba17782-3829-4de8-b49e-9eba858fc1ed","Type":"ContainerDied","Data":"b28cb91e8efc5938ac8b74a3d74189d2a6611eaf51fb4485ab2368fc6636aaee"} Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.896634 4992 generic.go:334] "Generic (PLEG): container finished" podID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerID="4053d7680d4a07f906b4cca19390824bbaa0af958fd22db3253dd80a46bb1868" exitCode=0 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.896710 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8cmz" event={"ID":"35ec2da1-7bca-4056-96eb-67a9687c4472","Type":"ContainerDied","Data":"4053d7680d4a07f906b4cca19390824bbaa0af958fd22db3253dd80a46bb1868"} Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.897835 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" event={"ID":"c8584f53-da6d-4606-8ac5-557d0fa24828","Type":"ContainerStarted","Data":"df83ddca6417d2381aac48839703e6accab0085bfcb7d1403d5ca29b61f632d6"} Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.899441 4992 generic.go:334] "Generic (PLEG): container finished" podID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerID="6ff7a09c2a3bccd4fbab24f05a4e6c0ec632dd893bd7ff90579f08d8f1e6803d" exitCode=0 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.899461 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grcdz" event={"ID":"acebc5f9-1793-4b93-a1e9-55604bb69996","Type":"ContainerDied","Data":"6ff7a09c2a3bccd4fbab24f05a4e6c0ec632dd893bd7ff90579f08d8f1e6803d"} Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.900849 4992 generic.go:334] "Generic (PLEG): container finished" podID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" containerID="27ead3d3f43bdede09c3827abb6f736a3996fe21b3dbc48dfa59cd63e6e2d29a" exitCode=0 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.900897 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" event={"ID":"ee4b8c43-5789-4d06-884e-f2fcc355eeec","Type":"ContainerDied","Data":"27ead3d3f43bdede09c3827abb6f736a3996fe21b3dbc48dfa59cd63e6e2d29a"} Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.902585 4992 generic.go:334] "Generic (PLEG): container finished" podID="151de3a2-278d-40d7-b869-023e9cd39273" containerID="7b2b6ab6417242dcf4b97a03d8a7a76f7d59029aa8d16535d2fbbac74da228b0" exitCode=0 Oct 04 04:39:33 crc kubenswrapper[4992]: I1004 04:39:33.902617 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w57r2" event={"ID":"151de3a2-278d-40d7-b869-023e9cd39273","Type":"ContainerDied","Data":"7b2b6ab6417242dcf4b97a03d8a7a76f7d59029aa8d16535d2fbbac74da228b0"} Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.182480 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.291756 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82qwm\" (UniqueName: \"kubernetes.io/projected/151de3a2-278d-40d7-b869-023e9cd39273-kube-api-access-82qwm\") pod \"151de3a2-278d-40d7-b869-023e9cd39273\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.292232 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-utilities\") pod \"151de3a2-278d-40d7-b869-023e9cd39273\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.292265 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-catalog-content\") pod \"151de3a2-278d-40d7-b869-023e9cd39273\" (UID: \"151de3a2-278d-40d7-b869-023e9cd39273\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.293607 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-utilities" (OuterVolumeSpecName: "utilities") pod "151de3a2-278d-40d7-b869-023e9cd39273" (UID: "151de3a2-278d-40d7-b869-023e9cd39273"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.296886 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.297852 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/151de3a2-278d-40d7-b869-023e9cd39273-kube-api-access-82qwm" (OuterVolumeSpecName: "kube-api-access-82qwm") pod "151de3a2-278d-40d7-b869-023e9cd39273" (UID: "151de3a2-278d-40d7-b869-023e9cd39273"). InnerVolumeSpecName "kube-api-access-82qwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.304614 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "151de3a2-278d-40d7-b869-023e9cd39273" (UID: "151de3a2-278d-40d7-b869-023e9cd39273"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.315169 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.319956 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.323258 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.383186 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397444 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-catalog-content\") pod \"35ec2da1-7bca-4056-96eb-67a9687c4472\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397497 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-utilities\") pod \"35ec2da1-7bca-4056-96eb-67a9687c4472\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397562 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjr97\" (UniqueName: \"kubernetes.io/projected/ee4b8c43-5789-4d06-884e-f2fcc355eeec-kube-api-access-hjr97\") pod \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397592 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-trusted-ca\") pod \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397616 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-utilities\") pod \"acebc5f9-1793-4b93-a1e9-55604bb69996\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397641 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-catalog-content\") pod \"acebc5f9-1793-4b93-a1e9-55604bb69996\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397667 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-operator-metrics\") pod \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\" (UID: \"ee4b8c43-5789-4d06-884e-f2fcc355eeec\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397697 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khdwl\" (UniqueName: \"kubernetes.io/projected/acebc5f9-1793-4b93-a1e9-55604bb69996-kube-api-access-khdwl\") pod \"acebc5f9-1793-4b93-a1e9-55604bb69996\" (UID: \"acebc5f9-1793-4b93-a1e9-55604bb69996\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397722 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ld6g\" (UniqueName: \"kubernetes.io/projected/35ec2da1-7bca-4056-96eb-67a9687c4472-kube-api-access-2ld6g\") pod \"35ec2da1-7bca-4056-96eb-67a9687c4472\" (UID: \"35ec2da1-7bca-4056-96eb-67a9687c4472\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397938 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/151de3a2-278d-40d7-b869-023e9cd39273-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.397952 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82qwm\" (UniqueName: \"kubernetes.io/projected/151de3a2-278d-40d7-b869-023e9cd39273-kube-api-access-82qwm\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.403633 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-utilities" (OuterVolumeSpecName: "utilities") pod "35ec2da1-7bca-4056-96eb-67a9687c4472" (UID: "35ec2da1-7bca-4056-96eb-67a9687c4472"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.405030 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ee4b8c43-5789-4d06-884e-f2fcc355eeec" (UID: "ee4b8c43-5789-4d06-884e-f2fcc355eeec"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.405038 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35ec2da1-7bca-4056-96eb-67a9687c4472-kube-api-access-2ld6g" (OuterVolumeSpecName: "kube-api-access-2ld6g") pod "35ec2da1-7bca-4056-96eb-67a9687c4472" (UID: "35ec2da1-7bca-4056-96eb-67a9687c4472"). InnerVolumeSpecName "kube-api-access-2ld6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.405713 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ee4b8c43-5789-4d06-884e-f2fcc355eeec" (UID: "ee4b8c43-5789-4d06-884e-f2fcc355eeec"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.407814 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acebc5f9-1793-4b93-a1e9-55604bb69996-kube-api-access-khdwl" (OuterVolumeSpecName: "kube-api-access-khdwl") pod "acebc5f9-1793-4b93-a1e9-55604bb69996" (UID: "acebc5f9-1793-4b93-a1e9-55604bb69996"). InnerVolumeSpecName "kube-api-access-khdwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.412890 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-utilities" (OuterVolumeSpecName: "utilities") pod "acebc5f9-1793-4b93-a1e9-55604bb69996" (UID: "acebc5f9-1793-4b93-a1e9-55604bb69996"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.413169 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee4b8c43-5789-4d06-884e-f2fcc355eeec-kube-api-access-hjr97" (OuterVolumeSpecName: "kube-api-access-hjr97") pod "ee4b8c43-5789-4d06-884e-f2fcc355eeec" (UID: "ee4b8c43-5789-4d06-884e-f2fcc355eeec"). InnerVolumeSpecName "kube-api-access-hjr97". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.488918 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35ec2da1-7bca-4056-96eb-67a9687c4472" (UID: "35ec2da1-7bca-4056-96eb-67a9687c4472"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.493613 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "acebc5f9-1793-4b93-a1e9-55604bb69996" (UID: "acebc5f9-1793-4b93-a1e9-55604bb69996"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498484 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-utilities\") pod \"cba17782-3829-4de8-b49e-9eba858fc1ed\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498520 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfgx4\" (UniqueName: \"kubernetes.io/projected/cba17782-3829-4de8-b49e-9eba858fc1ed-kube-api-access-tfgx4\") pod \"cba17782-3829-4de8-b49e-9eba858fc1ed\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498584 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-catalog-content\") pod \"cba17782-3829-4de8-b49e-9eba858fc1ed\" (UID: \"cba17782-3829-4de8-b49e-9eba858fc1ed\") " Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498773 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498786 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec2da1-7bca-4056-96eb-67a9687c4472-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498794 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjr97\" (UniqueName: \"kubernetes.io/projected/ee4b8c43-5789-4d06-884e-f2fcc355eeec-kube-api-access-hjr97\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498803 4992 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498813 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498821 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acebc5f9-1793-4b93-a1e9-55604bb69996-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498829 4992 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ee4b8c43-5789-4d06-884e-f2fcc355eeec-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498839 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khdwl\" (UniqueName: \"kubernetes.io/projected/acebc5f9-1793-4b93-a1e9-55604bb69996-kube-api-access-khdwl\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.498850 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ld6g\" (UniqueName: \"kubernetes.io/projected/35ec2da1-7bca-4056-96eb-67a9687c4472-kube-api-access-2ld6g\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.499172 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-utilities" (OuterVolumeSpecName: "utilities") pod "cba17782-3829-4de8-b49e-9eba858fc1ed" (UID: "cba17782-3829-4de8-b49e-9eba858fc1ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.510654 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba17782-3829-4de8-b49e-9eba858fc1ed-kube-api-access-tfgx4" (OuterVolumeSpecName: "kube-api-access-tfgx4") pod "cba17782-3829-4de8-b49e-9eba858fc1ed" (UID: "cba17782-3829-4de8-b49e-9eba858fc1ed"). InnerVolumeSpecName "kube-api-access-tfgx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.548181 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cba17782-3829-4de8-b49e-9eba858fc1ed" (UID: "cba17782-3829-4de8-b49e-9eba858fc1ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.600272 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.600752 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfgx4\" (UniqueName: \"kubernetes.io/projected/cba17782-3829-4de8-b49e-9eba858fc1ed-kube-api-access-tfgx4\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.601169 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cba17782-3829-4de8-b49e-9eba858fc1ed-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.911956 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grcdz" event={"ID":"acebc5f9-1793-4b93-a1e9-55604bb69996","Type":"ContainerDied","Data":"32e7d086e5a3f8ab09357a8d064a0565f24be01d6ba3587da26c926d8216ad06"} Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.912010 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grcdz" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.912023 4992 scope.go:117] "RemoveContainer" containerID="6ff7a09c2a3bccd4fbab24f05a4e6c0ec632dd893bd7ff90579f08d8f1e6803d" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.913576 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" event={"ID":"ee4b8c43-5789-4d06-884e-f2fcc355eeec","Type":"ContainerDied","Data":"f31a06270abf47b7476852f7b72a87ef19a2897f394eba9ac188538f806bbf14"} Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.913619 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tz4gp" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.916028 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w57r2" event={"ID":"151de3a2-278d-40d7-b869-023e9cd39273","Type":"ContainerDied","Data":"69f3de7cbb35e9e7f8ae2793e18a4f75bd2cfbadd58f3f400bc1b083a9158842"} Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.916142 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w57r2" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.920037 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4mz6" event={"ID":"cba17782-3829-4de8-b49e-9eba858fc1ed","Type":"ContainerDied","Data":"fcc012b1f36bef5c5a4540ec3120247695a361208cec40347cad6f81bc5c0625"} Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.920073 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4mz6" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.921824 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" event={"ID":"c8584f53-da6d-4606-8ac5-557d0fa24828","Type":"ContainerStarted","Data":"3228b824c970fc43c5a8c442347531307603b38499f080e54bf3f1161a21605b"} Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.922086 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.925264 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g8cmz" event={"ID":"35ec2da1-7bca-4056-96eb-67a9687c4472","Type":"ContainerDied","Data":"d8f66a612b2340b88f682d55763868b9933ed17f8a1f2b6f67592ea946ddac56"} Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.925428 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g8cmz" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.925612 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.933850 4992 scope.go:117] "RemoveContainer" containerID="a725dbfaf05d3ec48ee35f147e381f25e00bd3c97c33d107f6b1a4c1308341c6" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.946714 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ztwc8" podStartSLOduration=1.946671493 podStartE2EDuration="1.946671493s" podCreationTimestamp="2025-10-04 04:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:39:34.94326927 +0000 UTC m=+408.790944778" watchObservedRunningTime="2025-10-04 04:39:34.946671493 +0000 UTC m=+408.794346961" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.959194 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w57r2"] Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.965484 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w57r2"] Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.968620 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tz4gp"] Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.968837 4992 scope.go:117] "RemoveContainer" containerID="1105c2532fb9c9ea1da96965d86dd4dee35e07081bab24811c2896c3162f2b14" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.973106 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tz4gp"] Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.975466 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-grcdz"] Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.984025 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-grcdz"] Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.987295 4992 scope.go:117] "RemoveContainer" containerID="27ead3d3f43bdede09c3827abb6f736a3996fe21b3dbc48dfa59cd63e6e2d29a" Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.989480 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4mz6"] Oct 04 04:39:34 crc kubenswrapper[4992]: I1004 04:39:34.992661 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c4mz6"] Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.007560 4992 scope.go:117] "RemoveContainer" containerID="7b2b6ab6417242dcf4b97a03d8a7a76f7d59029aa8d16535d2fbbac74da228b0" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.024661 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g8cmz"] Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.027611 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g8cmz"] Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.027685 4992 scope.go:117] "RemoveContainer" containerID="062904d058ef5df016ad2a0c91e9341e5bd388133c89cc7b49c2e312a9ca45e2" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.040855 4992 scope.go:117] "RemoveContainer" containerID="0bebeb8ce526839ccafa25c11e8d56b8d8ff2f443b1b79dbe684dc7ec0a24801" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.052965 4992 scope.go:117] "RemoveContainer" containerID="b28cb91e8efc5938ac8b74a3d74189d2a6611eaf51fb4485ab2368fc6636aaee" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.065410 4992 scope.go:117] "RemoveContainer" containerID="85c6ac5f60583a2db4bfe93391c94f84b8e9bbb74b11cc1d8a7bcac4803047e6" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.078977 4992 scope.go:117] "RemoveContainer" containerID="2f63cbc106580d36e0c6f107171833104935d71765c8f80b75f6b14b987b610c" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.095173 4992 scope.go:117] "RemoveContainer" containerID="4053d7680d4a07f906b4cca19390824bbaa0af958fd22db3253dd80a46bb1868" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.111947 4992 scope.go:117] "RemoveContainer" containerID="4b0f9c9fe1d8f451ce8fb31d569ed91ab3c2bfdc6413d51bf2c2d93d0ebf613b" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.125436 4992 scope.go:117] "RemoveContainer" containerID="ea07cd786f44516980c9679d4dbd0f08f692235b18a98a71e5312fc6849aceab" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543056 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s55dk"] Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543664 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="151de3a2-278d-40d7-b869-023e9cd39273" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543676 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="151de3a2-278d-40d7-b869-023e9cd39273" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543686 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="151de3a2-278d-40d7-b869-023e9cd39273" containerName="extract-content" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543692 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="151de3a2-278d-40d7-b869-023e9cd39273" containerName="extract-content" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543703 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerName="extract-utilities" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543710 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerName="extract-utilities" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543741 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerName="extract-content" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543747 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerName="extract-content" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543756 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerName="extract-utilities" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543761 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerName="extract-utilities" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543771 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerName="extract-content" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543777 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerName="extract-content" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543784 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerName="extract-utilities" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543790 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerName="extract-utilities" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543797 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543803 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543813 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543819 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543825 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" containerName="marketplace-operator" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543831 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" containerName="marketplace-operator" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543839 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543844 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543851 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="151de3a2-278d-40d7-b869-023e9cd39273" containerName="extract-utilities" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543857 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="151de3a2-278d-40d7-b869-023e9cd39273" containerName="extract-utilities" Oct 04 04:39:35 crc kubenswrapper[4992]: E1004 04:39:35.543866 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerName="extract-content" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543871 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerName="extract-content" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543963 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" containerName="marketplace-operator" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543975 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543983 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="151de3a2-278d-40d7-b869-023e9cd39273" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543991 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.543999 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" containerName="registry-server" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.544653 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.546443 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.562021 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s55dk"] Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.712053 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-utilities\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.712101 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-catalog-content\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.712173 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmxc2\" (UniqueName: \"kubernetes.io/projected/e208c780-5970-45e0-adce-55980aa4bd88-kube-api-access-kmxc2\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.813317 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-catalog-content\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.813411 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmxc2\" (UniqueName: \"kubernetes.io/projected/e208c780-5970-45e0-adce-55980aa4bd88-kube-api-access-kmxc2\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.813453 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-utilities\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.813850 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-catalog-content\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.813897 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-utilities\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.829238 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmxc2\" (UniqueName: \"kubernetes.io/projected/e208c780-5970-45e0-adce-55980aa4bd88-kube-api-access-kmxc2\") pod \"community-operators-s55dk\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:35 crc kubenswrapper[4992]: I1004 04:39:35.881945 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.049297 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s55dk"] Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.326107 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="151de3a2-278d-40d7-b869-023e9cd39273" path="/var/lib/kubelet/pods/151de3a2-278d-40d7-b869-023e9cd39273/volumes" Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.327006 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35ec2da1-7bca-4056-96eb-67a9687c4472" path="/var/lib/kubelet/pods/35ec2da1-7bca-4056-96eb-67a9687c4472/volumes" Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.327736 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acebc5f9-1793-4b93-a1e9-55604bb69996" path="/var/lib/kubelet/pods/acebc5f9-1793-4b93-a1e9-55604bb69996/volumes" Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.329035 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cba17782-3829-4de8-b49e-9eba858fc1ed" path="/var/lib/kubelet/pods/cba17782-3829-4de8-b49e-9eba858fc1ed/volumes" Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.329866 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee4b8c43-5789-4d06-884e-f2fcc355eeec" path="/var/lib/kubelet/pods/ee4b8c43-5789-4d06-884e-f2fcc355eeec/volumes" Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.949921 4992 generic.go:334] "Generic (PLEG): container finished" podID="e208c780-5970-45e0-adce-55980aa4bd88" containerID="7a586454bc1eba450665cc764f4f77352a045cdedc5a8b9d0b56936893033642" exitCode=0 Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.950789 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s55dk" event={"ID":"e208c780-5970-45e0-adce-55980aa4bd88","Type":"ContainerDied","Data":"7a586454bc1eba450665cc764f4f77352a045cdedc5a8b9d0b56936893033642"} Oct 04 04:39:36 crc kubenswrapper[4992]: I1004 04:39:36.950814 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s55dk" event={"ID":"e208c780-5970-45e0-adce-55980aa4bd88","Type":"ContainerStarted","Data":"b9dac6345f7b1747198c4c15d0e7f3fe002a6ec4f557f4b15ef2c61c0d6037b6"} Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.358196 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lp25k"] Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.360258 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.360643 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lp25k"] Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.361950 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.435346 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plm9b\" (UniqueName: \"kubernetes.io/projected/166adf65-fdf7-422e-b3a8-43ff22d793a5-kube-api-access-plm9b\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.435797 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/166adf65-fdf7-422e-b3a8-43ff22d793a5-utilities\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.435885 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/166adf65-fdf7-422e-b3a8-43ff22d793a5-catalog-content\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.537348 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plm9b\" (UniqueName: \"kubernetes.io/projected/166adf65-fdf7-422e-b3a8-43ff22d793a5-kube-api-access-plm9b\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.537460 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/166adf65-fdf7-422e-b3a8-43ff22d793a5-utilities\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.537510 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/166adf65-fdf7-422e-b3a8-43ff22d793a5-catalog-content\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.538213 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/166adf65-fdf7-422e-b3a8-43ff22d793a5-catalog-content\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.538336 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/166adf65-fdf7-422e-b3a8-43ff22d793a5-utilities\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.558546 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plm9b\" (UniqueName: \"kubernetes.io/projected/166adf65-fdf7-422e-b3a8-43ff22d793a5-kube-api-access-plm9b\") pod \"certified-operators-lp25k\" (UID: \"166adf65-fdf7-422e-b3a8-43ff22d793a5\") " pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.681516 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.945622 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mq7hd"] Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.946906 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.950793 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 04:39:37 crc kubenswrapper[4992]: I1004 04:39:37.954243 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mq7hd"] Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.044775 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-utilities\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.044849 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpgrr\" (UniqueName: \"kubernetes.io/projected/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-kube-api-access-wpgrr\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.044873 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-catalog-content\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.119306 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lp25k"] Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.145944 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-utilities\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.146025 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpgrr\" (UniqueName: \"kubernetes.io/projected/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-kube-api-access-wpgrr\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.146060 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-catalog-content\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.146456 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-utilities\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.146560 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-catalog-content\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.170346 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpgrr\" (UniqueName: \"kubernetes.io/projected/c0186a4f-5104-4beb-a51e-5aec49ed8f0c-kube-api-access-wpgrr\") pod \"redhat-operators-mq7hd\" (UID: \"c0186a4f-5104-4beb-a51e-5aec49ed8f0c\") " pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.268269 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:38 crc kubenswrapper[4992]: I1004 04:39:38.959785 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lp25k" event={"ID":"166adf65-fdf7-422e-b3a8-43ff22d793a5","Type":"ContainerStarted","Data":"a963e64cc1e297a65103beaaf5b4a09c12be5ac25f84d83808e4e964be122cb3"} Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.599987 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mq7hd"] Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.745619 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p2k7x"] Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.746916 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.752507 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.754680 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2k7x"] Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.870843 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d9fr\" (UniqueName: \"kubernetes.io/projected/36aa6d66-c182-47cb-8899-588cbe55f588-kube-api-access-2d9fr\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.870883 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36aa6d66-c182-47cb-8899-588cbe55f588-catalog-content\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.870946 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36aa6d66-c182-47cb-8899-588cbe55f588-utilities\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.964779 4992 generic.go:334] "Generic (PLEG): container finished" podID="e208c780-5970-45e0-adce-55980aa4bd88" containerID="fee0f62e350126841a91f76687a950b6dc2338a7c839d67cbfca66afa568f2ae" exitCode=0 Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.964852 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s55dk" event={"ID":"e208c780-5970-45e0-adce-55980aa4bd88","Type":"ContainerDied","Data":"fee0f62e350126841a91f76687a950b6dc2338a7c839d67cbfca66afa568f2ae"} Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.965860 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mq7hd" event={"ID":"c0186a4f-5104-4beb-a51e-5aec49ed8f0c","Type":"ContainerStarted","Data":"8b632d9d8caa2cb38f249f59989a43159f37d8c4249b0b2070e4d8d9b730c879"} Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.971226 4992 generic.go:334] "Generic (PLEG): container finished" podID="166adf65-fdf7-422e-b3a8-43ff22d793a5" containerID="26782382f7796d810063334e0bbe8b98784ccc9ba1a91e4774d39e0a6ac8c049" exitCode=0 Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.971339 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lp25k" event={"ID":"166adf65-fdf7-422e-b3a8-43ff22d793a5","Type":"ContainerDied","Data":"26782382f7796d810063334e0bbe8b98784ccc9ba1a91e4774d39e0a6ac8c049"} Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.971747 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d9fr\" (UniqueName: \"kubernetes.io/projected/36aa6d66-c182-47cb-8899-588cbe55f588-kube-api-access-2d9fr\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.971805 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36aa6d66-c182-47cb-8899-588cbe55f588-catalog-content\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.971924 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36aa6d66-c182-47cb-8899-588cbe55f588-utilities\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.972671 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36aa6d66-c182-47cb-8899-588cbe55f588-catalog-content\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.973015 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36aa6d66-c182-47cb-8899-588cbe55f588-utilities\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:39 crc kubenswrapper[4992]: I1004 04:39:39.998668 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d9fr\" (UniqueName: \"kubernetes.io/projected/36aa6d66-c182-47cb-8899-588cbe55f588-kube-api-access-2d9fr\") pod \"redhat-marketplace-p2k7x\" (UID: \"36aa6d66-c182-47cb-8899-588cbe55f588\") " pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:40 crc kubenswrapper[4992]: I1004 04:39:40.097198 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:40 crc kubenswrapper[4992]: I1004 04:39:40.326833 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2k7x"] Oct 04 04:39:40 crc kubenswrapper[4992]: W1004 04:39:40.362825 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36aa6d66_c182_47cb_8899_588cbe55f588.slice/crio-01966d94447c7f78a98cda82859602880cacbff1ef69dfe49d6a152274f80c4a WatchSource:0}: Error finding container 01966d94447c7f78a98cda82859602880cacbff1ef69dfe49d6a152274f80c4a: Status 404 returned error can't find the container with id 01966d94447c7f78a98cda82859602880cacbff1ef69dfe49d6a152274f80c4a Oct 04 04:39:40 crc kubenswrapper[4992]: I1004 04:39:40.978475 4992 generic.go:334] "Generic (PLEG): container finished" podID="36aa6d66-c182-47cb-8899-588cbe55f588" containerID="c2d0d82caaa93a44ad1771c2a79514117288d746d086da4bd43a4440c9414b65" exitCode=0 Oct 04 04:39:40 crc kubenswrapper[4992]: I1004 04:39:40.978541 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2k7x" event={"ID":"36aa6d66-c182-47cb-8899-588cbe55f588","Type":"ContainerDied","Data":"c2d0d82caaa93a44ad1771c2a79514117288d746d086da4bd43a4440c9414b65"} Oct 04 04:39:40 crc kubenswrapper[4992]: I1004 04:39:40.978570 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2k7x" event={"ID":"36aa6d66-c182-47cb-8899-588cbe55f588","Type":"ContainerStarted","Data":"01966d94447c7f78a98cda82859602880cacbff1ef69dfe49d6a152274f80c4a"} Oct 04 04:39:40 crc kubenswrapper[4992]: I1004 04:39:40.984113 4992 generic.go:334] "Generic (PLEG): container finished" podID="c0186a4f-5104-4beb-a51e-5aec49ed8f0c" containerID="ab706c8be7ca27864a34aea16c775285864cdc6953ce6668b98fcd25fce28d2b" exitCode=0 Oct 04 04:39:40 crc kubenswrapper[4992]: I1004 04:39:40.984150 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mq7hd" event={"ID":"c0186a4f-5104-4beb-a51e-5aec49ed8f0c","Type":"ContainerDied","Data":"ab706c8be7ca27864a34aea16c775285864cdc6953ce6668b98fcd25fce28d2b"} Oct 04 04:39:43 crc kubenswrapper[4992]: I1004 04:39:42.999855 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s55dk" event={"ID":"e208c780-5970-45e0-adce-55980aa4bd88","Type":"ContainerStarted","Data":"89d646e67c891a5fb50ea641c0ae6168148cf2c34cc8cbc665bd62ce609f59c9"} Oct 04 04:39:43 crc kubenswrapper[4992]: I1004 04:39:43.019255 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s55dk" podStartSLOduration=3.466731346 podStartE2EDuration="8.019231574s" podCreationTimestamp="2025-10-04 04:39:35 +0000 UTC" firstStartedPulling="2025-10-04 04:39:36.953049064 +0000 UTC m=+410.800724532" lastFinishedPulling="2025-10-04 04:39:41.505549292 +0000 UTC m=+415.353224760" observedRunningTime="2025-10-04 04:39:43.014964977 +0000 UTC m=+416.862640445" watchObservedRunningTime="2025-10-04 04:39:43.019231574 +0000 UTC m=+416.866907052" Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.010731 4992 generic.go:334] "Generic (PLEG): container finished" podID="36aa6d66-c182-47cb-8899-588cbe55f588" containerID="6fb80c0c56d9c39b08bca51ef1877bdb2f039177e03983f1e530946bd857ed69" exitCode=0 Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.010830 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2k7x" event={"ID":"36aa6d66-c182-47cb-8899-588cbe55f588","Type":"ContainerDied","Data":"6fb80c0c56d9c39b08bca51ef1877bdb2f039177e03983f1e530946bd857ed69"} Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.012887 4992 generic.go:334] "Generic (PLEG): container finished" podID="c0186a4f-5104-4beb-a51e-5aec49ed8f0c" containerID="9eae54559167db7ecb1531880490085550588030dc25f9e6fd1bf6e2a47d8c3c" exitCode=0 Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.012957 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mq7hd" event={"ID":"c0186a4f-5104-4beb-a51e-5aec49ed8f0c","Type":"ContainerDied","Data":"9eae54559167db7ecb1531880490085550588030dc25f9e6fd1bf6e2a47d8c3c"} Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.016326 4992 generic.go:334] "Generic (PLEG): container finished" podID="166adf65-fdf7-422e-b3a8-43ff22d793a5" containerID="8a54700d6dc61edca7984a789f204362ae1974e2737f4c20c796d15f2a9129f1" exitCode=0 Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.016426 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lp25k" event={"ID":"166adf65-fdf7-422e-b3a8-43ff22d793a5","Type":"ContainerDied","Data":"8a54700d6dc61edca7984a789f204362ae1974e2737f4c20c796d15f2a9129f1"} Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.882850 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.883165 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:45 crc kubenswrapper[4992]: I1004 04:39:45.924230 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:47 crc kubenswrapper[4992]: I1004 04:39:47.037165 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lp25k" event={"ID":"166adf65-fdf7-422e-b3a8-43ff22d793a5","Type":"ContainerStarted","Data":"7cbc8a020e21e1aa0948bf5665fe8927a95622fc3fa27793b14eef4f9212a800"} Oct 04 04:39:47 crc kubenswrapper[4992]: I1004 04:39:47.054299 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lp25k" podStartSLOduration=3.387774716 podStartE2EDuration="10.054284616s" podCreationTimestamp="2025-10-04 04:39:37 +0000 UTC" firstStartedPulling="2025-10-04 04:39:39.972782788 +0000 UTC m=+413.820458256" lastFinishedPulling="2025-10-04 04:39:46.639292688 +0000 UTC m=+420.486968156" observedRunningTime="2025-10-04 04:39:47.05333649 +0000 UTC m=+420.901011968" watchObservedRunningTime="2025-10-04 04:39:47.054284616 +0000 UTC m=+420.901960084" Oct 04 04:39:47 crc kubenswrapper[4992]: I1004 04:39:47.682065 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:47 crc kubenswrapper[4992]: I1004 04:39:47.682111 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:48 crc kubenswrapper[4992]: I1004 04:39:48.044598 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2k7x" event={"ID":"36aa6d66-c182-47cb-8899-588cbe55f588","Type":"ContainerStarted","Data":"a371d4fdb6bc563602ca710f0fa0404e1f1fb26b0f1d3968f54132d2f05dde07"} Oct 04 04:39:48 crc kubenswrapper[4992]: I1004 04:39:48.046761 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mq7hd" event={"ID":"c0186a4f-5104-4beb-a51e-5aec49ed8f0c","Type":"ContainerStarted","Data":"f0f57e328a51bdfffc8973dbfc6b81e912afd563722f7d961bb867dfbdecde06"} Oct 04 04:39:48 crc kubenswrapper[4992]: I1004 04:39:48.065918 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mq7hd" podStartSLOduration=4.8454130410000005 podStartE2EDuration="11.065903202s" podCreationTimestamp="2025-10-04 04:39:37 +0000 UTC" firstStartedPulling="2025-10-04 04:39:40.985450523 +0000 UTC m=+414.833125991" lastFinishedPulling="2025-10-04 04:39:47.205940684 +0000 UTC m=+421.053616152" observedRunningTime="2025-10-04 04:39:48.062622922 +0000 UTC m=+421.910298450" watchObservedRunningTime="2025-10-04 04:39:48.065903202 +0000 UTC m=+421.913578670" Oct 04 04:39:48 crc kubenswrapper[4992]: I1004 04:39:48.268730 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:48 crc kubenswrapper[4992]: I1004 04:39:48.268816 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:48 crc kubenswrapper[4992]: I1004 04:39:48.724301 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lp25k" podUID="166adf65-fdf7-422e-b3a8-43ff22d793a5" containerName="registry-server" probeResult="failure" output=< Oct 04 04:39:48 crc kubenswrapper[4992]: timeout: failed to connect service ":50051" within 1s Oct 04 04:39:48 crc kubenswrapper[4992]: > Oct 04 04:39:49 crc kubenswrapper[4992]: I1004 04:39:49.067083 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p2k7x" podStartSLOduration=3.316113485 podStartE2EDuration="10.067063741s" podCreationTimestamp="2025-10-04 04:39:39 +0000 UTC" firstStartedPulling="2025-10-04 04:39:40.98171075 +0000 UTC m=+414.829386218" lastFinishedPulling="2025-10-04 04:39:47.732661006 +0000 UTC m=+421.580336474" observedRunningTime="2025-10-04 04:39:49.065683623 +0000 UTC m=+422.913359091" watchObservedRunningTime="2025-10-04 04:39:49.067063741 +0000 UTC m=+422.914739209" Oct 04 04:39:49 crc kubenswrapper[4992]: I1004 04:39:49.304796 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mq7hd" podUID="c0186a4f-5104-4beb-a51e-5aec49ed8f0c" containerName="registry-server" probeResult="failure" output=< Oct 04 04:39:49 crc kubenswrapper[4992]: timeout: failed to connect service ":50051" within 1s Oct 04 04:39:49 crc kubenswrapper[4992]: > Oct 04 04:39:50 crc kubenswrapper[4992]: I1004 04:39:50.098015 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:50 crc kubenswrapper[4992]: I1004 04:39:50.098081 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:50 crc kubenswrapper[4992]: I1004 04:39:50.164034 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:39:50 crc kubenswrapper[4992]: I1004 04:39:50.894744 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:39:50 crc kubenswrapper[4992]: I1004 04:39:50.894843 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:39:55 crc kubenswrapper[4992]: I1004 04:39:55.920726 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s55dk" Oct 04 04:39:57 crc kubenswrapper[4992]: I1004 04:39:57.720457 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:57 crc kubenswrapper[4992]: I1004 04:39:57.759062 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lp25k" Oct 04 04:39:58 crc kubenswrapper[4992]: I1004 04:39:58.334498 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:39:58 crc kubenswrapper[4992]: I1004 04:39:58.379610 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mq7hd" Oct 04 04:40:00 crc kubenswrapper[4992]: I1004 04:40:00.142813 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p2k7x" Oct 04 04:40:20 crc kubenswrapper[4992]: I1004 04:40:20.894436 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:40:20 crc kubenswrapper[4992]: I1004 04:40:20.894877 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:40:20 crc kubenswrapper[4992]: I1004 04:40:20.894926 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:40:20 crc kubenswrapper[4992]: I1004 04:40:20.895524 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93092393a842d35831bfc9be98a84a2f8a76369bad06835685afb28ae1178e48"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:40:20 crc kubenswrapper[4992]: I1004 04:40:20.895576 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://93092393a842d35831bfc9be98a84a2f8a76369bad06835685afb28ae1178e48" gracePeriod=600 Oct 04 04:40:22 crc kubenswrapper[4992]: I1004 04:40:22.247237 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="93092393a842d35831bfc9be98a84a2f8a76369bad06835685afb28ae1178e48" exitCode=0 Oct 04 04:40:22 crc kubenswrapper[4992]: I1004 04:40:22.247415 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"93092393a842d35831bfc9be98a84a2f8a76369bad06835685afb28ae1178e48"} Oct 04 04:40:22 crc kubenswrapper[4992]: I1004 04:40:22.247880 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"34bbae1fd997a7077d97cc7f2a26b45479ea4683901678b60f3f46c1a59664b1"} Oct 04 04:40:22 crc kubenswrapper[4992]: I1004 04:40:22.247911 4992 scope.go:117] "RemoveContainer" containerID="f83df1db0bf1f5d86a8d2d60217a982ffdab0eb8f7d7fde1f4995f1b7db259ef" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.361903 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xd98n"] Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.363788 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.388105 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xd98n"] Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.395548 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-registry-tls\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.395828 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/884e51d7-6f2a-4091-9e96-740be05a0281-registry-certificates\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.395936 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.396040 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5x4r\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-kube-api-access-h5x4r\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.396131 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/884e51d7-6f2a-4091-9e96-740be05a0281-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.396252 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/884e51d7-6f2a-4091-9e96-740be05a0281-trusted-ca\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.396381 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/884e51d7-6f2a-4091-9e96-740be05a0281-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.396465 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-bound-sa-token\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.431107 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.497422 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/884e51d7-6f2a-4091-9e96-740be05a0281-registry-certificates\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.497719 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5x4r\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-kube-api-access-h5x4r\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.497822 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/884e51d7-6f2a-4091-9e96-740be05a0281-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.497935 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/884e51d7-6f2a-4091-9e96-740be05a0281-trusted-ca\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.498052 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/884e51d7-6f2a-4091-9e96-740be05a0281-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.498217 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-bound-sa-token\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.498342 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-registry-tls\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.499256 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/884e51d7-6f2a-4091-9e96-740be05a0281-trusted-ca\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.499349 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/884e51d7-6f2a-4091-9e96-740be05a0281-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.499416 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/884e51d7-6f2a-4091-9e96-740be05a0281-registry-certificates\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.504009 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/884e51d7-6f2a-4091-9e96-740be05a0281-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.507202 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-registry-tls\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.514349 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-bound-sa-token\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.515424 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5x4r\" (UniqueName: \"kubernetes.io/projected/884e51d7-6f2a-4091-9e96-740be05a0281-kube-api-access-h5x4r\") pod \"image-registry-66df7c8f76-xd98n\" (UID: \"884e51d7-6f2a-4091-9e96-740be05a0281\") " pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:34 crc kubenswrapper[4992]: I1004 04:40:34.680713 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:35 crc kubenswrapper[4992]: I1004 04:40:35.107232 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xd98n"] Oct 04 04:40:35 crc kubenswrapper[4992]: I1004 04:40:35.333748 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" event={"ID":"884e51d7-6f2a-4091-9e96-740be05a0281","Type":"ContainerStarted","Data":"617ef1cddf9275a428741ccdec4fdf11053f55c35cb91db644276d800b335608"} Oct 04 04:40:36 crc kubenswrapper[4992]: I1004 04:40:36.340867 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" event={"ID":"884e51d7-6f2a-4091-9e96-740be05a0281","Type":"ContainerStarted","Data":"beffb794c5558595e625d8ccd797c4d30ce4348b45c5372233037570c0e9d65a"} Oct 04 04:40:36 crc kubenswrapper[4992]: I1004 04:40:36.341042 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:36 crc kubenswrapper[4992]: I1004 04:40:36.364221 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" podStartSLOduration=2.364189427 podStartE2EDuration="2.364189427s" podCreationTimestamp="2025-10-04 04:40:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:40:36.362648986 +0000 UTC m=+470.210324474" watchObservedRunningTime="2025-10-04 04:40:36.364189427 +0000 UTC m=+470.211864935" Oct 04 04:40:54 crc kubenswrapper[4992]: I1004 04:40:54.685434 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-xd98n" Oct 04 04:40:54 crc kubenswrapper[4992]: I1004 04:40:54.738706 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fz68t"] Oct 04 04:41:19 crc kubenswrapper[4992]: I1004 04:41:19.778369 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" podUID="78372347-a30d-4a0f-a4d4-c92a1999eaad" containerName="registry" containerID="cri-o://de510d584b716716a3dbc0eb8972939ce14fb13130706471106edce880eca207" gracePeriod=30 Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.001302 4992 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-fz68t container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.20:5000/healthz\": dial tcp 10.217.0.20:5000: connect: connection refused" start-of-body= Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.001404 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" podUID="78372347-a30d-4a0f-a4d4-c92a1999eaad" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.20:5000/healthz\": dial tcp 10.217.0.20:5000: connect: connection refused" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.580682 4992 generic.go:334] "Generic (PLEG): container finished" podID="78372347-a30d-4a0f-a4d4-c92a1999eaad" containerID="de510d584b716716a3dbc0eb8972939ce14fb13130706471106edce880eca207" exitCode=0 Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.580868 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" event={"ID":"78372347-a30d-4a0f-a4d4-c92a1999eaad","Type":"ContainerDied","Data":"de510d584b716716a3dbc0eb8972939ce14fb13130706471106edce880eca207"} Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.581033 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" event={"ID":"78372347-a30d-4a0f-a4d4-c92a1999eaad","Type":"ContainerDied","Data":"0751b6b8404873485d3c9ce2df880f8943e6fb368d74adf09a839b8d8f4ca9fb"} Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.581052 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0751b6b8404873485d3c9ce2df880f8943e6fb368d74adf09a839b8d8f4ca9fb" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.609090 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.627851 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-bound-sa-token\") pod \"78372347-a30d-4a0f-a4d4-c92a1999eaad\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.627926 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-tls\") pod \"78372347-a30d-4a0f-a4d4-c92a1999eaad\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.627951 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78372347-a30d-4a0f-a4d4-c92a1999eaad-ca-trust-extracted\") pod \"78372347-a30d-4a0f-a4d4-c92a1999eaad\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.628000 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glx8g\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-kube-api-access-glx8g\") pod \"78372347-a30d-4a0f-a4d4-c92a1999eaad\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.628067 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-trusted-ca\") pod \"78372347-a30d-4a0f-a4d4-c92a1999eaad\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.628090 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78372347-a30d-4a0f-a4d4-c92a1999eaad-installation-pull-secrets\") pod \"78372347-a30d-4a0f-a4d4-c92a1999eaad\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.628148 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-certificates\") pod \"78372347-a30d-4a0f-a4d4-c92a1999eaad\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.628383 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"78372347-a30d-4a0f-a4d4-c92a1999eaad\" (UID: \"78372347-a30d-4a0f-a4d4-c92a1999eaad\") " Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.631733 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "78372347-a30d-4a0f-a4d4-c92a1999eaad" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.633162 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "78372347-a30d-4a0f-a4d4-c92a1999eaad" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.637419 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78372347-a30d-4a0f-a4d4-c92a1999eaad-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "78372347-a30d-4a0f-a4d4-c92a1999eaad" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.637557 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-kube-api-access-glx8g" (OuterVolumeSpecName: "kube-api-access-glx8g") pod "78372347-a30d-4a0f-a4d4-c92a1999eaad" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad"). InnerVolumeSpecName "kube-api-access-glx8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.637929 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "78372347-a30d-4a0f-a4d4-c92a1999eaad" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.638702 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "78372347-a30d-4a0f-a4d4-c92a1999eaad" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.649584 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "78372347-a30d-4a0f-a4d4-c92a1999eaad" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.650947 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78372347-a30d-4a0f-a4d4-c92a1999eaad-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "78372347-a30d-4a0f-a4d4-c92a1999eaad" (UID: "78372347-a30d-4a0f-a4d4-c92a1999eaad"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.729376 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.729411 4992 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78372347-a30d-4a0f-a4d4-c92a1999eaad-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.729421 4992 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.729430 4992 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.729440 4992 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.729449 4992 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78372347-a30d-4a0f-a4d4-c92a1999eaad-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:20 crc kubenswrapper[4992]: I1004 04:41:20.729457 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glx8g\" (UniqueName: \"kubernetes.io/projected/78372347-a30d-4a0f-a4d4-c92a1999eaad-kube-api-access-glx8g\") on node \"crc\" DevicePath \"\"" Oct 04 04:41:21 crc kubenswrapper[4992]: I1004 04:41:21.585805 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fz68t" Oct 04 04:41:21 crc kubenswrapper[4992]: I1004 04:41:21.629949 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fz68t"] Oct 04 04:41:21 crc kubenswrapper[4992]: I1004 04:41:21.633224 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fz68t"] Oct 04 04:41:22 crc kubenswrapper[4992]: I1004 04:41:22.329956 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78372347-a30d-4a0f-a4d4-c92a1999eaad" path="/var/lib/kubelet/pods/78372347-a30d-4a0f-a4d4-c92a1999eaad/volumes" Oct 04 04:41:58 crc kubenswrapper[4992]: I1004 04:41:58.024776 4992 scope.go:117] "RemoveContainer" containerID="de510d584b716716a3dbc0eb8972939ce14fb13130706471106edce880eca207" Oct 04 04:42:50 crc kubenswrapper[4992]: I1004 04:42:50.895194 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:42:50 crc kubenswrapper[4992]: I1004 04:42:50.896250 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:43:20 crc kubenswrapper[4992]: I1004 04:43:20.894118 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:43:20 crc kubenswrapper[4992]: I1004 04:43:20.894675 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:43:50 crc kubenswrapper[4992]: I1004 04:43:50.894930 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:43:50 crc kubenswrapper[4992]: I1004 04:43:50.895911 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:43:50 crc kubenswrapper[4992]: I1004 04:43:50.896043 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:43:50 crc kubenswrapper[4992]: I1004 04:43:50.897322 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34bbae1fd997a7077d97cc7f2a26b45479ea4683901678b60f3f46c1a59664b1"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:43:50 crc kubenswrapper[4992]: I1004 04:43:50.897617 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://34bbae1fd997a7077d97cc7f2a26b45479ea4683901678b60f3f46c1a59664b1" gracePeriod=600 Oct 04 04:43:51 crc kubenswrapper[4992]: I1004 04:43:51.359380 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="34bbae1fd997a7077d97cc7f2a26b45479ea4683901678b60f3f46c1a59664b1" exitCode=0 Oct 04 04:43:51 crc kubenswrapper[4992]: I1004 04:43:51.359552 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"34bbae1fd997a7077d97cc7f2a26b45479ea4683901678b60f3f46c1a59664b1"} Oct 04 04:43:51 crc kubenswrapper[4992]: I1004 04:43:51.359778 4992 scope.go:117] "RemoveContainer" containerID="93092393a842d35831bfc9be98a84a2f8a76369bad06835685afb28ae1178e48" Oct 04 04:43:52 crc kubenswrapper[4992]: I1004 04:43:52.367903 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"391734b7432560d7fffa75f76956def8a70e770707e3c21215c4ea365ca34204"} Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.135278 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66"] Oct 04 04:45:00 crc kubenswrapper[4992]: E1004 04:45:00.136086 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78372347-a30d-4a0f-a4d4-c92a1999eaad" containerName="registry" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.136103 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="78372347-a30d-4a0f-a4d4-c92a1999eaad" containerName="registry" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.136223 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="78372347-a30d-4a0f-a4d4-c92a1999eaad" containerName="registry" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.136675 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.138873 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.138892 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.145857 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66"] Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.158794 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5994ffa6-01dd-4b1c-b592-be6f6845f20a-secret-volume\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.159087 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5994ffa6-01dd-4b1c-b592-be6f6845f20a-config-volume\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.159239 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvt28\" (UniqueName: \"kubernetes.io/projected/5994ffa6-01dd-4b1c-b592-be6f6845f20a-kube-api-access-kvt28\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.261040 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5994ffa6-01dd-4b1c-b592-be6f6845f20a-secret-volume\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.261130 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5994ffa6-01dd-4b1c-b592-be6f6845f20a-config-volume\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.261178 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvt28\" (UniqueName: \"kubernetes.io/projected/5994ffa6-01dd-4b1c-b592-be6f6845f20a-kube-api-access-kvt28\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.262225 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5994ffa6-01dd-4b1c-b592-be6f6845f20a-config-volume\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.267259 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5994ffa6-01dd-4b1c-b592-be6f6845f20a-secret-volume\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.279499 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvt28\" (UniqueName: \"kubernetes.io/projected/5994ffa6-01dd-4b1c-b592-be6f6845f20a-kube-api-access-kvt28\") pod \"collect-profiles-29325885-87l66\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.513891 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.703179 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66"] Oct 04 04:45:00 crc kubenswrapper[4992]: I1004 04:45:00.740939 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" event={"ID":"5994ffa6-01dd-4b1c-b592-be6f6845f20a","Type":"ContainerStarted","Data":"394d2da2b51ed576f3ab5b77d7fa2dba0ecac02efa8e76debf63b98f5bb290db"} Oct 04 04:45:01 crc kubenswrapper[4992]: I1004 04:45:01.748866 4992 generic.go:334] "Generic (PLEG): container finished" podID="5994ffa6-01dd-4b1c-b592-be6f6845f20a" containerID="2bd0d4cfad0092c0b1428d6cd51b868fb9da1bf1286b48a137db4334e46c5a1e" exitCode=0 Oct 04 04:45:01 crc kubenswrapper[4992]: I1004 04:45:01.748919 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" event={"ID":"5994ffa6-01dd-4b1c-b592-be6f6845f20a","Type":"ContainerDied","Data":"2bd0d4cfad0092c0b1428d6cd51b868fb9da1bf1286b48a137db4334e46c5a1e"} Oct 04 04:45:02 crc kubenswrapper[4992]: I1004 04:45:02.954516 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:02 crc kubenswrapper[4992]: I1004 04:45:02.988970 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvt28\" (UniqueName: \"kubernetes.io/projected/5994ffa6-01dd-4b1c-b592-be6f6845f20a-kube-api-access-kvt28\") pod \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " Oct 04 04:45:02 crc kubenswrapper[4992]: I1004 04:45:02.989028 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5994ffa6-01dd-4b1c-b592-be6f6845f20a-config-volume\") pod \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " Oct 04 04:45:02 crc kubenswrapper[4992]: I1004 04:45:02.989053 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5994ffa6-01dd-4b1c-b592-be6f6845f20a-secret-volume\") pod \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\" (UID: \"5994ffa6-01dd-4b1c-b592-be6f6845f20a\") " Oct 04 04:45:02 crc kubenswrapper[4992]: I1004 04:45:02.989674 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5994ffa6-01dd-4b1c-b592-be6f6845f20a-config-volume" (OuterVolumeSpecName: "config-volume") pod "5994ffa6-01dd-4b1c-b592-be6f6845f20a" (UID: "5994ffa6-01dd-4b1c-b592-be6f6845f20a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:02 crc kubenswrapper[4992]: I1004 04:45:02.993320 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5994ffa6-01dd-4b1c-b592-be6f6845f20a-kube-api-access-kvt28" (OuterVolumeSpecName: "kube-api-access-kvt28") pod "5994ffa6-01dd-4b1c-b592-be6f6845f20a" (UID: "5994ffa6-01dd-4b1c-b592-be6f6845f20a"). InnerVolumeSpecName "kube-api-access-kvt28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:45:02 crc kubenswrapper[4992]: I1004 04:45:02.993435 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5994ffa6-01dd-4b1c-b592-be6f6845f20a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5994ffa6-01dd-4b1c-b592-be6f6845f20a" (UID: "5994ffa6-01dd-4b1c-b592-be6f6845f20a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:03 crc kubenswrapper[4992]: I1004 04:45:03.089660 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvt28\" (UniqueName: \"kubernetes.io/projected/5994ffa6-01dd-4b1c-b592-be6f6845f20a-kube-api-access-kvt28\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:03 crc kubenswrapper[4992]: I1004 04:45:03.089695 4992 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5994ffa6-01dd-4b1c-b592-be6f6845f20a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:03 crc kubenswrapper[4992]: I1004 04:45:03.089706 4992 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5994ffa6-01dd-4b1c-b592-be6f6845f20a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:03 crc kubenswrapper[4992]: I1004 04:45:03.759184 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" event={"ID":"5994ffa6-01dd-4b1c-b592-be6f6845f20a","Type":"ContainerDied","Data":"394d2da2b51ed576f3ab5b77d7fa2dba0ecac02efa8e76debf63b98f5bb290db"} Oct 04 04:45:03 crc kubenswrapper[4992]: I1004 04:45:03.759232 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="394d2da2b51ed576f3ab5b77d7fa2dba0ecac02efa8e76debf63b98f5bb290db" Oct 04 04:45:03 crc kubenswrapper[4992]: I1004 04:45:03.759242 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.041607 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fb7m9"] Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.043058 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" podUID="d8c6ed95-6bcb-4a2d-b7d4-de1419159770" containerName="controller-manager" containerID="cri-o://a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df" gracePeriod=30 Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.178018 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8"] Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.178226 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" podUID="4256fad3-f7d5-4441-a4f0-ee84712d7321" containerName="route-controller-manager" containerID="cri-o://9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e" gracePeriod=30 Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.444623 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.501069 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-client-ca\") pod \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.501122 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-proxy-ca-bundles\") pod \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.502975 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-client-ca" (OuterVolumeSpecName: "client-ca") pod "d8c6ed95-6bcb-4a2d-b7d4-de1419159770" (UID: "d8c6ed95-6bcb-4a2d-b7d4-de1419159770"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.502994 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d8c6ed95-6bcb-4a2d-b7d4-de1419159770" (UID: "d8c6ed95-6bcb-4a2d-b7d4-de1419159770"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.534088 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.601807 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-serving-cert\") pod \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.601855 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-config\") pod \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.601937 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6877b\" (UniqueName: \"kubernetes.io/projected/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-kube-api-access-6877b\") pod \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\" (UID: \"d8c6ed95-6bcb-4a2d-b7d4-de1419159770\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.602157 4992 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.602172 4992 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.602589 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-config" (OuterVolumeSpecName: "config") pod "d8c6ed95-6bcb-4a2d-b7d4-de1419159770" (UID: "d8c6ed95-6bcb-4a2d-b7d4-de1419159770"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.606623 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d8c6ed95-6bcb-4a2d-b7d4-de1419159770" (UID: "d8c6ed95-6bcb-4a2d-b7d4-de1419159770"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.606734 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-kube-api-access-6877b" (OuterVolumeSpecName: "kube-api-access-6877b") pod "d8c6ed95-6bcb-4a2d-b7d4-de1419159770" (UID: "d8c6ed95-6bcb-4a2d-b7d4-de1419159770"). InnerVolumeSpecName "kube-api-access-6877b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.702908 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config\") pod \"4256fad3-f7d5-4441-a4f0-ee84712d7321\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.702986 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz22k\" (UniqueName: \"kubernetes.io/projected/4256fad3-f7d5-4441-a4f0-ee84712d7321-kube-api-access-zz22k\") pod \"4256fad3-f7d5-4441-a4f0-ee84712d7321\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.703015 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-client-ca\") pod \"4256fad3-f7d5-4441-a4f0-ee84712d7321\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.703106 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4256fad3-f7d5-4441-a4f0-ee84712d7321-serving-cert\") pod \"4256fad3-f7d5-4441-a4f0-ee84712d7321\" (UID: \"4256fad3-f7d5-4441-a4f0-ee84712d7321\") " Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.703342 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.703374 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.703387 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6877b\" (UniqueName: \"kubernetes.io/projected/d8c6ed95-6bcb-4a2d-b7d4-de1419159770-kube-api-access-6877b\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.704245 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-client-ca" (OuterVolumeSpecName: "client-ca") pod "4256fad3-f7d5-4441-a4f0-ee84712d7321" (UID: "4256fad3-f7d5-4441-a4f0-ee84712d7321"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.704317 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config" (OuterVolumeSpecName: "config") pod "4256fad3-f7d5-4441-a4f0-ee84712d7321" (UID: "4256fad3-f7d5-4441-a4f0-ee84712d7321"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.706762 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4256fad3-f7d5-4441-a4f0-ee84712d7321-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4256fad3-f7d5-4441-a4f0-ee84712d7321" (UID: "4256fad3-f7d5-4441-a4f0-ee84712d7321"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.707219 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4256fad3-f7d5-4441-a4f0-ee84712d7321-kube-api-access-zz22k" (OuterVolumeSpecName: "kube-api-access-zz22k") pod "4256fad3-f7d5-4441-a4f0-ee84712d7321" (UID: "4256fad3-f7d5-4441-a4f0-ee84712d7321"). InnerVolumeSpecName "kube-api-access-zz22k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.805194 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.805240 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz22k\" (UniqueName: \"kubernetes.io/projected/4256fad3-f7d5-4441-a4f0-ee84712d7321-kube-api-access-zz22k\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.805255 4992 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4256fad3-f7d5-4441-a4f0-ee84712d7321-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.805267 4992 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4256fad3-f7d5-4441-a4f0-ee84712d7321-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.889634 4992 generic.go:334] "Generic (PLEG): container finished" podID="4256fad3-f7d5-4441-a4f0-ee84712d7321" containerID="9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e" exitCode=0 Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.889709 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.889739 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" event={"ID":"4256fad3-f7d5-4441-a4f0-ee84712d7321","Type":"ContainerDied","Data":"9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e"} Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.889789 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8" event={"ID":"4256fad3-f7d5-4441-a4f0-ee84712d7321","Type":"ContainerDied","Data":"2d17d269865c95134c47ae932655df0ca28b9147f303af4098d5bb20d70cab59"} Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.889813 4992 scope.go:117] "RemoveContainer" containerID="9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.890958 4992 generic.go:334] "Generic (PLEG): container finished" podID="d8c6ed95-6bcb-4a2d-b7d4-de1419159770" containerID="a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df" exitCode=0 Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.890984 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" event={"ID":"d8c6ed95-6bcb-4a2d-b7d4-de1419159770","Type":"ContainerDied","Data":"a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df"} Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.891009 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" event={"ID":"d8c6ed95-6bcb-4a2d-b7d4-de1419159770","Type":"ContainerDied","Data":"f98ab7d9a4d88da5bd4ebf6e77dc6435ff8659e2e3c10efc6c482ec64b6bfd76"} Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.891015 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fb7m9" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.903617 4992 scope.go:117] "RemoveContainer" containerID="9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e" Oct 04 04:45:27 crc kubenswrapper[4992]: E1004 04:45:27.904176 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e\": container with ID starting with 9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e not found: ID does not exist" containerID="9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.904205 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e"} err="failed to get container status \"9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e\": rpc error: code = NotFound desc = could not find container \"9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e\": container with ID starting with 9006b48dc9d09e0c87df830f609d984ea292eb4ad9e43a6c1406bfdddcf7f13e not found: ID does not exist" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.904224 4992 scope.go:117] "RemoveContainer" containerID="a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.915985 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8"] Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.917986 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4tqf8"] Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.923026 4992 scope.go:117] "RemoveContainer" containerID="a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df" Oct 04 04:45:27 crc kubenswrapper[4992]: E1004 04:45:27.923549 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df\": container with ID starting with a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df not found: ID does not exist" containerID="a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.923585 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df"} err="failed to get container status \"a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df\": rpc error: code = NotFound desc = could not find container \"a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df\": container with ID starting with a6877e7db3780be5fe04516bb61bd41a2a56542f6d07c3c2a38be7861fb554df not found: ID does not exist" Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.928208 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fb7m9"] Oct 04 04:45:27 crc kubenswrapper[4992]: I1004 04:45:27.930000 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fb7m9"] Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.326876 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4256fad3-f7d5-4441-a4f0-ee84712d7321" path="/var/lib/kubelet/pods/4256fad3-f7d5-4441-a4f0-ee84712d7321/volumes" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.327695 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8c6ed95-6bcb-4a2d-b7d4-de1419159770" path="/var/lib/kubelet/pods/d8c6ed95-6bcb-4a2d-b7d4-de1419159770/volumes" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.566696 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4"] Oct 04 04:45:28 crc kubenswrapper[4992]: E1004 04:45:28.566920 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c6ed95-6bcb-4a2d-b7d4-de1419159770" containerName="controller-manager" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.566941 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c6ed95-6bcb-4a2d-b7d4-de1419159770" containerName="controller-manager" Oct 04 04:45:28 crc kubenswrapper[4992]: E1004 04:45:28.566965 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5994ffa6-01dd-4b1c-b592-be6f6845f20a" containerName="collect-profiles" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.566975 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="5994ffa6-01dd-4b1c-b592-be6f6845f20a" containerName="collect-profiles" Oct 04 04:45:28 crc kubenswrapper[4992]: E1004 04:45:28.566989 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4256fad3-f7d5-4441-a4f0-ee84712d7321" containerName="route-controller-manager" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.566998 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="4256fad3-f7d5-4441-a4f0-ee84712d7321" containerName="route-controller-manager" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.567099 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="5994ffa6-01dd-4b1c-b592-be6f6845f20a" containerName="collect-profiles" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.567113 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="4256fad3-f7d5-4441-a4f0-ee84712d7321" containerName="route-controller-manager" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.567130 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c6ed95-6bcb-4a2d-b7d4-de1419159770" containerName="controller-manager" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.570103 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.573310 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.573552 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.573734 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.573885 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.574166 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.574060 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.576114 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-79c45cfd86-7mxnk"] Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.577129 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.579135 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.579651 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.579944 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.580276 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.580530 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.581247 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.584384 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79c45cfd86-7mxnk"] Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.586802 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4"] Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.592297 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614286 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f65e99a-e267-4c51-a156-88b72bc552dc-serving-cert\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614336 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-serving-cert\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614414 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-config\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614505 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnsgh\" (UniqueName: \"kubernetes.io/projected/2f65e99a-e267-4c51-a156-88b72bc552dc-kube-api-access-qnsgh\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614565 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-client-ca\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614602 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-proxy-ca-bundles\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614786 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-config\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614858 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-client-ca\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.614906 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9rzs\" (UniqueName: \"kubernetes.io/projected/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-kube-api-access-j9rzs\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715487 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnsgh\" (UniqueName: \"kubernetes.io/projected/2f65e99a-e267-4c51-a156-88b72bc552dc-kube-api-access-qnsgh\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715536 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-client-ca\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715553 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-proxy-ca-bundles\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715587 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-config\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715602 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-client-ca\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715627 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9rzs\" (UniqueName: \"kubernetes.io/projected/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-kube-api-access-j9rzs\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715651 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f65e99a-e267-4c51-a156-88b72bc552dc-serving-cert\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715668 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-serving-cert\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.715684 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-config\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.716671 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-client-ca\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.716699 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-client-ca\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.717076 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-proxy-ca-bundles\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.717447 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-config\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.717714 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f65e99a-e267-4c51-a156-88b72bc552dc-config\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.723331 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f65e99a-e267-4c51-a156-88b72bc552dc-serving-cert\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.723517 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-serving-cert\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.732971 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9rzs\" (UniqueName: \"kubernetes.io/projected/aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9-kube-api-access-j9rzs\") pod \"route-controller-manager-6dfb67454-gkxf4\" (UID: \"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9\") " pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.742473 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnsgh\" (UniqueName: \"kubernetes.io/projected/2f65e99a-e267-4c51-a156-88b72bc552dc-kube-api-access-qnsgh\") pod \"controller-manager-79c45cfd86-7mxnk\" (UID: \"2f65e99a-e267-4c51-a156-88b72bc552dc\") " pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.898611 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:28 crc kubenswrapper[4992]: I1004 04:45:28.911878 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.079164 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79c45cfd86-7mxnk"] Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.125735 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4"] Oct 04 04:45:29 crc kubenswrapper[4992]: W1004 04:45:29.134923 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa4eb5af_23d2_44bb_aa2c_5638faa9fdb9.slice/crio-b87e9253043361d3657ea672f915cdcbbe8d708baebf938ee37c3b679b1301a6 WatchSource:0}: Error finding container b87e9253043361d3657ea672f915cdcbbe8d708baebf938ee37c3b679b1301a6: Status 404 returned error can't find the container with id b87e9253043361d3657ea672f915cdcbbe8d708baebf938ee37c3b679b1301a6 Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.904668 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" event={"ID":"2f65e99a-e267-4c51-a156-88b72bc552dc","Type":"ContainerStarted","Data":"2fb5a57a1c7f033c4a638d6de8bb88064d87490792fadfa16f73950eb9d861f5"} Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.905005 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.905021 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" event={"ID":"2f65e99a-e267-4c51-a156-88b72bc552dc","Type":"ContainerStarted","Data":"490b8cc91820e2d6fe2f703346351b5ec24f2a6076171c1edf5100876f5f3ee3"} Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.906046 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" event={"ID":"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9","Type":"ContainerStarted","Data":"a3ec0d6f15849b5c7f1210247032f79e7bfc84154bac800038b80f84095dd2bf"} Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.906086 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" event={"ID":"aa4eb5af-23d2-44bb-aa2c-5638faa9fdb9","Type":"ContainerStarted","Data":"b87e9253043361d3657ea672f915cdcbbe8d708baebf938ee37c3b679b1301a6"} Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.906446 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.909457 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.910479 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.923201 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-79c45cfd86-7mxnk" podStartSLOduration=2.923184562 podStartE2EDuration="2.923184562s" podCreationTimestamp="2025-10-04 04:45:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:45:29.918839376 +0000 UTC m=+763.766514864" watchObservedRunningTime="2025-10-04 04:45:29.923184562 +0000 UTC m=+763.770860030" Oct 04 04:45:29 crc kubenswrapper[4992]: I1004 04:45:29.941239 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6dfb67454-gkxf4" podStartSLOduration=2.941223052 podStartE2EDuration="2.941223052s" podCreationTimestamp="2025-10-04 04:45:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:45:29.940254946 +0000 UTC m=+763.787930414" watchObservedRunningTime="2025-10-04 04:45:29.941223052 +0000 UTC m=+763.788898520" Oct 04 04:45:35 crc kubenswrapper[4992]: I1004 04:45:35.619754 4992 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.054028 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4gjdf"] Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.057126 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4gjdf" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.060333 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7tqhx"] Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.066093 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4gjdf"] Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.066184 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7tqhx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.067483 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.068851 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.068960 4992 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-f7hdn" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.069241 4992 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-sb4r6" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.070118 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7tqhx"] Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.085955 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-5lblx"] Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.089055 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.090829 4992 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-999px" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.098308 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-5lblx"] Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.190095 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvxhw\" (UniqueName: \"kubernetes.io/projected/059e879c-9cec-4e22-8702-8c5f93cffd5d-kube-api-access-xvxhw\") pod \"cert-manager-5b446d88c5-7tqhx\" (UID: \"059e879c-9cec-4e22-8702-8c5f93cffd5d\") " pod="cert-manager/cert-manager-5b446d88c5-7tqhx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.190500 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksvsk\" (UniqueName: \"kubernetes.io/projected/911792e6-61de-4469-9ced-16ffca1da0e7-kube-api-access-ksvsk\") pod \"cert-manager-webhook-5655c58dd6-5lblx\" (UID: \"911792e6-61de-4469-9ced-16ffca1da0e7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.190551 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74d6j\" (UniqueName: \"kubernetes.io/projected/eefe44db-fa85-476c-9900-7c1e3d46a63f-kube-api-access-74d6j\") pod \"cert-manager-cainjector-7f985d654d-4gjdf\" (UID: \"eefe44db-fa85-476c-9900-7c1e3d46a63f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4gjdf" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.291560 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvxhw\" (UniqueName: \"kubernetes.io/projected/059e879c-9cec-4e22-8702-8c5f93cffd5d-kube-api-access-xvxhw\") pod \"cert-manager-5b446d88c5-7tqhx\" (UID: \"059e879c-9cec-4e22-8702-8c5f93cffd5d\") " pod="cert-manager/cert-manager-5b446d88c5-7tqhx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.291620 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksvsk\" (UniqueName: \"kubernetes.io/projected/911792e6-61de-4469-9ced-16ffca1da0e7-kube-api-access-ksvsk\") pod \"cert-manager-webhook-5655c58dd6-5lblx\" (UID: \"911792e6-61de-4469-9ced-16ffca1da0e7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.291660 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74d6j\" (UniqueName: \"kubernetes.io/projected/eefe44db-fa85-476c-9900-7c1e3d46a63f-kube-api-access-74d6j\") pod \"cert-manager-cainjector-7f985d654d-4gjdf\" (UID: \"eefe44db-fa85-476c-9900-7c1e3d46a63f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4gjdf" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.309525 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksvsk\" (UniqueName: \"kubernetes.io/projected/911792e6-61de-4469-9ced-16ffca1da0e7-kube-api-access-ksvsk\") pod \"cert-manager-webhook-5655c58dd6-5lblx\" (UID: \"911792e6-61de-4469-9ced-16ffca1da0e7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.310296 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvxhw\" (UniqueName: \"kubernetes.io/projected/059e879c-9cec-4e22-8702-8c5f93cffd5d-kube-api-access-xvxhw\") pod \"cert-manager-5b446d88c5-7tqhx\" (UID: \"059e879c-9cec-4e22-8702-8c5f93cffd5d\") " pod="cert-manager/cert-manager-5b446d88c5-7tqhx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.314977 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74d6j\" (UniqueName: \"kubernetes.io/projected/eefe44db-fa85-476c-9900-7c1e3d46a63f-kube-api-access-74d6j\") pod \"cert-manager-cainjector-7f985d654d-4gjdf\" (UID: \"eefe44db-fa85-476c-9900-7c1e3d46a63f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4gjdf" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.394704 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4gjdf" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.413703 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7tqhx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.420024 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.830531 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4gjdf"] Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.834273 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7tqhx"] Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.839344 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:46:19 crc kubenswrapper[4992]: I1004 04:46:19.916213 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-5lblx"] Oct 04 04:46:19 crc kubenswrapper[4992]: W1004 04:46:19.922658 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod911792e6_61de_4469_9ced_16ffca1da0e7.slice/crio-6f0d3075a6c386e56d38d5459f34e7d520b4045cbe32f06fbad3e0ab90196fdf WatchSource:0}: Error finding container 6f0d3075a6c386e56d38d5459f34e7d520b4045cbe32f06fbad3e0ab90196fdf: Status 404 returned error can't find the container with id 6f0d3075a6c386e56d38d5459f34e7d520b4045cbe32f06fbad3e0ab90196fdf Oct 04 04:46:20 crc kubenswrapper[4992]: I1004 04:46:20.197497 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4gjdf" event={"ID":"eefe44db-fa85-476c-9900-7c1e3d46a63f","Type":"ContainerStarted","Data":"7eb1735db0346220e0e2c4215655942ed2f0542925cff3a0a51f6547970c3522"} Oct 04 04:46:20 crc kubenswrapper[4992]: I1004 04:46:20.198651 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7tqhx" event={"ID":"059e879c-9cec-4e22-8702-8c5f93cffd5d","Type":"ContainerStarted","Data":"5e3dd3c3a3a43869e3629772b68d51306a63877ca44c30904c142d80098a623f"} Oct 04 04:46:20 crc kubenswrapper[4992]: I1004 04:46:20.199757 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" event={"ID":"911792e6-61de-4469-9ced-16ffca1da0e7","Type":"ContainerStarted","Data":"6f0d3075a6c386e56d38d5459f34e7d520b4045cbe32f06fbad3e0ab90196fdf"} Oct 04 04:46:20 crc kubenswrapper[4992]: I1004 04:46:20.894728 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:46:20 crc kubenswrapper[4992]: I1004 04:46:20.894818 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:46:24 crc kubenswrapper[4992]: I1004 04:46:24.223477 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4gjdf" event={"ID":"eefe44db-fa85-476c-9900-7c1e3d46a63f","Type":"ContainerStarted","Data":"2acb1dc065fe052a91d94a7455032d7e0eb45348cc80aa663e62ce54f3146626"} Oct 04 04:46:24 crc kubenswrapper[4992]: I1004 04:46:24.226039 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7tqhx" event={"ID":"059e879c-9cec-4e22-8702-8c5f93cffd5d","Type":"ContainerStarted","Data":"a7f274540746c01c674ade470932aad5fa97ce206c801718d798e0b8dbe36490"} Oct 04 04:46:24 crc kubenswrapper[4992]: I1004 04:46:24.227521 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" event={"ID":"911792e6-61de-4469-9ced-16ffca1da0e7","Type":"ContainerStarted","Data":"01551a8d76b46c13c8a1a89566857cc0843d7052ca104b4266f222d5434f8adb"} Oct 04 04:46:24 crc kubenswrapper[4992]: I1004 04:46:24.227961 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" Oct 04 04:46:24 crc kubenswrapper[4992]: I1004 04:46:24.243136 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-4gjdf" podStartSLOduration=1.72572367 podStartE2EDuration="5.243117022s" podCreationTimestamp="2025-10-04 04:46:19 +0000 UTC" firstStartedPulling="2025-10-04 04:46:19.83916931 +0000 UTC m=+813.686844778" lastFinishedPulling="2025-10-04 04:46:23.356562672 +0000 UTC m=+817.204238130" observedRunningTime="2025-10-04 04:46:24.240829671 +0000 UTC m=+818.088505159" watchObservedRunningTime="2025-10-04 04:46:24.243117022 +0000 UTC m=+818.090792500" Oct 04 04:46:24 crc kubenswrapper[4992]: I1004 04:46:24.261295 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-7tqhx" podStartSLOduration=1.737147564 podStartE2EDuration="5.261272655s" podCreationTimestamp="2025-10-04 04:46:19 +0000 UTC" firstStartedPulling="2025-10-04 04:46:19.840861885 +0000 UTC m=+813.688537353" lastFinishedPulling="2025-10-04 04:46:23.364986976 +0000 UTC m=+817.212662444" observedRunningTime="2025-10-04 04:46:24.25920775 +0000 UTC m=+818.106883228" watchObservedRunningTime="2025-10-04 04:46:24.261272655 +0000 UTC m=+818.108948133" Oct 04 04:46:24 crc kubenswrapper[4992]: I1004 04:46:24.281821 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" podStartSLOduration=1.7486271 podStartE2EDuration="5.281801322s" podCreationTimestamp="2025-10-04 04:46:19 +0000 UTC" firstStartedPulling="2025-10-04 04:46:19.928514588 +0000 UTC m=+813.776190056" lastFinishedPulling="2025-10-04 04:46:23.46168881 +0000 UTC m=+817.309364278" observedRunningTime="2025-10-04 04:46:24.280931419 +0000 UTC m=+818.128606907" watchObservedRunningTime="2025-10-04 04:46:24.281801322 +0000 UTC m=+818.129476790" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.424436 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-5lblx" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.452636 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fnr76"] Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.453274 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovn-controller" containerID="cri-o://24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940" gracePeriod=30 Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.453385 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="nbdb" containerID="cri-o://f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f" gracePeriod=30 Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.453507 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="northd" containerID="cri-o://24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f" gracePeriod=30 Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.453559 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc" gracePeriod=30 Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.453599 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kube-rbac-proxy-node" containerID="cri-o://9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c" gracePeriod=30 Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.453634 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovn-acl-logging" containerID="cri-o://db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c" gracePeriod=30 Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.453875 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="sbdb" containerID="cri-o://46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9" gracePeriod=30 Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.525945 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" containerID="cri-o://80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720" gracePeriod=30 Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.823706 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/3.log" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.828130 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovn-acl-logging/0.log" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.828883 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovn-controller/0.log" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.829865 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894274 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pg9kr"] Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894570 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894587 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894602 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="northd" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894611 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="northd" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894620 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="nbdb" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894677 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="nbdb" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894689 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894698 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894709 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894717 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894728 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovn-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894736 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovn-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894751 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovn-acl-logging" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894759 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovn-acl-logging" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894772 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="sbdb" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894784 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="sbdb" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894795 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894805 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894818 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kube-rbac-proxy-node" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894827 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kube-rbac-proxy-node" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894840 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894849 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894862 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894870 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: E1004 04:46:29.894880 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kubecfg-setup" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.894890 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kubecfg-setup" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895015 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovn-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895028 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895038 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895048 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895058 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895069 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="nbdb" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895081 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895092 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="sbdb" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895103 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovn-acl-logging" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895114 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="northd" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895127 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="kube-rbac-proxy-node" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.895350 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerName="ovnkube-controller" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.897592 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.935925 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-ovn\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936031 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-log-socket\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936060 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-openvswitch\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936077 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936136 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936104 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-env-overrides\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936152 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-log-socket" (OuterVolumeSpecName: "log-socket") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936516 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-bin\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936558 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-systemd\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936577 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-slash\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936593 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936599 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-systemd-units\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936628 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936623 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936642 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936657 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-slash" (OuterVolumeSpecName: "host-slash") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936680 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-ovn-kubernetes\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936709 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-netd\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936718 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936733 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936742 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936736 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-script-lib\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936795 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-etc-openvswitch\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936834 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovn-node-metrics-cert\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936854 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936868 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-config\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936888 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-kubelet\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936927 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7qm9\" (UniqueName: \"kubernetes.io/projected/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-kube-api-access-b7qm9\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936948 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-node-log\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936967 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-var-lib-openvswitch\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.936995 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-netns\") pod \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\" (UID: \"2e301ed8-9506-4d6e-bbce-d2ad3817ca36\") " Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937177 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-kubelet\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937212 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-run-netns\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937236 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-cni-bin\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937238 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937247 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937260 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937280 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937285 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937286 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-ovn\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937299 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-node-log" (OuterVolumeSpecName: "node-log") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937315 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937391 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937410 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovn-node-metrics-cert\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937438 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-node-log\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937460 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-systemd\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937479 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-systemd-units\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937501 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-log-socket\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937546 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-run-ovn-kubernetes\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937586 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-slash\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937608 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovnkube-script-lib\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937630 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tvk6\" (UniqueName: \"kubernetes.io/projected/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-kube-api-access-9tvk6\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937654 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-var-lib-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937679 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-cni-netd\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937700 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-etc-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937725 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-env-overrides\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937751 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovnkube-config\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937808 4992 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937824 4992 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937837 4992 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937849 4992 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937863 4992 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937876 4992 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937887 4992 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937898 4992 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937909 4992 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937920 4992 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937935 4992 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937947 4992 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937958 4992 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937970 4992 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937982 4992 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.937995 4992 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.938007 4992 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.942868 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.943310 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-kube-api-access-b7qm9" (OuterVolumeSpecName: "kube-api-access-b7qm9") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "kube-api-access-b7qm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:46:29 crc kubenswrapper[4992]: I1004 04:46:29.950406 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "2e301ed8-9506-4d6e-bbce-d2ad3817ca36" (UID: "2e301ed8-9506-4d6e-bbce-d2ad3817ca36"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039086 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039167 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovn-node-metrics-cert\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039205 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-node-log\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039246 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-systemd\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039274 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-node-log\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039289 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-systemd-units\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039237 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039296 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-systemd\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039386 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-systemd-units\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039413 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-log-socket\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039462 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-run-ovn-kubernetes\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039493 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-slash\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039509 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-log-socket\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039517 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovnkube-script-lib\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039553 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-slash\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039518 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-run-ovn-kubernetes\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039577 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tvk6\" (UniqueName: \"kubernetes.io/projected/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-kube-api-access-9tvk6\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039608 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-var-lib-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039637 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-cni-netd\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039661 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-etc-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039677 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-var-lib-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039690 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovnkube-config\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039697 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-cni-netd\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039718 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-etc-openvswitch\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039718 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-env-overrides\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039781 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-kubelet\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039821 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-run-netns\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039851 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-cni-bin\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039880 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039919 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-ovn\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039924 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-run-netns\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039960 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-run-ovn\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.039986 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-kubelet\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.040052 4992 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.040012 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-cni-bin\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.040075 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7qm9\" (UniqueName: \"kubernetes.io/projected/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-kube-api-access-b7qm9\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.040114 4992 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2e301ed8-9506-4d6e-bbce-d2ad3817ca36-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.040036 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.040615 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovnkube-config\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.040680 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovnkube-script-lib\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.040699 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-env-overrides\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.043477 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-ovn-node-metrics-cert\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.055642 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tvk6\" (UniqueName: \"kubernetes.io/projected/b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22-kube-api-access-9tvk6\") pod \"ovnkube-node-pg9kr\" (UID: \"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22\") " pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.211035 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.264825 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/2.log" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.265949 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/1.log" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.266017 4992 generic.go:334] "Generic (PLEG): container finished" podID="2b6879c5-46f4-4d87-959b-03dff08cff4c" containerID="f5c27d0d56f2bb47591a59be7ae33f559c7fefcb65ab4b8a0ea45e4c3b25a61c" exitCode=2 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.266091 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhq64" event={"ID":"2b6879c5-46f4-4d87-959b-03dff08cff4c","Type":"ContainerDied","Data":"f5c27d0d56f2bb47591a59be7ae33f559c7fefcb65ab4b8a0ea45e4c3b25a61c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.266130 4992 scope.go:117] "RemoveContainer" containerID="e9f3cb5ecf25bcac5ab719bf1fb801fdc9083338426e4e368ddcd4d298d59105" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.266815 4992 scope.go:117] "RemoveContainer" containerID="f5c27d0d56f2bb47591a59be7ae33f559c7fefcb65ab4b8a0ea45e4c3b25a61c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.269131 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovnkube-controller/3.log" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.275112 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovn-acl-logging/0.log" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.277421 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fnr76_2e301ed8-9506-4d6e-bbce-d2ad3817ca36/ovn-controller/0.log" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278092 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720" exitCode=0 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278132 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9" exitCode=0 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278144 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f" exitCode=0 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278154 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f" exitCode=0 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278162 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc" exitCode=0 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278174 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c" exitCode=0 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278185 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c" exitCode=143 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278195 4992 generic.go:334] "Generic (PLEG): container finished" podID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" containerID="24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940" exitCode=143 Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278293 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278343 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278385 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278401 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278413 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278457 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278451 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278472 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278636 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278649 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278657 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278665 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278673 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278679 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278685 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278693 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278704 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278719 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278733 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278745 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278754 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278763 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278768 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278774 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278779 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278787 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278796 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278801 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278811 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278828 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278839 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278848 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278856 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278862 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278871 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278881 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278890 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278898 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278905 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278913 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fnr76" event={"ID":"2e301ed8-9506-4d6e-bbce-d2ad3817ca36","Type":"ContainerDied","Data":"2976e96f9c7962bd0e54bc0651abe8f495b61bf46ef033a09c179306afea1e9c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278921 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278928 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278938 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278944 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278949 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278954 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278960 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278965 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278972 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.278977 4992 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.280095 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"f6a70f5659ac13147be6e90e9053fb9daeb1fe7ed1f323318cc80f1a1e4002f5"} Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.350741 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fnr76"] Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.354564 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fnr76"] Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.380554 4992 scope.go:117] "RemoveContainer" containerID="80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.408820 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.494064 4992 scope.go:117] "RemoveContainer" containerID="46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.514968 4992 scope.go:117] "RemoveContainer" containerID="f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.527418 4992 scope.go:117] "RemoveContainer" containerID="24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.538941 4992 scope.go:117] "RemoveContainer" containerID="6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.555342 4992 scope.go:117] "RemoveContainer" containerID="9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.571587 4992 scope.go:117] "RemoveContainer" containerID="db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.585335 4992 scope.go:117] "RemoveContainer" containerID="24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.609002 4992 scope.go:117] "RemoveContainer" containerID="93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.786423 4992 scope.go:117] "RemoveContainer" containerID="80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.788115 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": container with ID starting with 80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720 not found: ID does not exist" containerID="80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.788158 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} err="failed to get container status \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": rpc error: code = NotFound desc = could not find container \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": container with ID starting with 80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.788182 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.788725 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": container with ID starting with bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7 not found: ID does not exist" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.788820 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} err="failed to get container status \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": rpc error: code = NotFound desc = could not find container \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": container with ID starting with bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.788860 4992 scope.go:117] "RemoveContainer" containerID="46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.789495 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": container with ID starting with 46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9 not found: ID does not exist" containerID="46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.789573 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} err="failed to get container status \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": rpc error: code = NotFound desc = could not find container \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": container with ID starting with 46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.789627 4992 scope.go:117] "RemoveContainer" containerID="f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.790247 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": container with ID starting with f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f not found: ID does not exist" containerID="f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.790285 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} err="failed to get container status \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": rpc error: code = NotFound desc = could not find container \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": container with ID starting with f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.790311 4992 scope.go:117] "RemoveContainer" containerID="24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.790764 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": container with ID starting with 24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f not found: ID does not exist" containerID="24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.790810 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} err="failed to get container status \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": rpc error: code = NotFound desc = could not find container \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": container with ID starting with 24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.790838 4992 scope.go:117] "RemoveContainer" containerID="6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.791155 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": container with ID starting with 6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc not found: ID does not exist" containerID="6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.791208 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} err="failed to get container status \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": rpc error: code = NotFound desc = could not find container \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": container with ID starting with 6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.791244 4992 scope.go:117] "RemoveContainer" containerID="9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.791872 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": container with ID starting with 9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c not found: ID does not exist" containerID="9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.791898 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} err="failed to get container status \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": rpc error: code = NotFound desc = could not find container \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": container with ID starting with 9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.791918 4992 scope.go:117] "RemoveContainer" containerID="db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.792760 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": container with ID starting with db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c not found: ID does not exist" containerID="db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.792796 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} err="failed to get container status \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": rpc error: code = NotFound desc = could not find container \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": container with ID starting with db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.792815 4992 scope.go:117] "RemoveContainer" containerID="24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.793119 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": container with ID starting with 24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940 not found: ID does not exist" containerID="24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.793167 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} err="failed to get container status \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": rpc error: code = NotFound desc = could not find container \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": container with ID starting with 24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.793193 4992 scope.go:117] "RemoveContainer" containerID="93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3" Oct 04 04:46:30 crc kubenswrapper[4992]: E1004 04:46:30.793590 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": container with ID starting with 93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3 not found: ID does not exist" containerID="93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.793614 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} err="failed to get container status \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": rpc error: code = NotFound desc = could not find container \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": container with ID starting with 93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.793629 4992 scope.go:117] "RemoveContainer" containerID="80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.794038 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} err="failed to get container status \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": rpc error: code = NotFound desc = could not find container \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": container with ID starting with 80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.794067 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.794452 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} err="failed to get container status \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": rpc error: code = NotFound desc = could not find container \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": container with ID starting with bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.794476 4992 scope.go:117] "RemoveContainer" containerID="46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.794780 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} err="failed to get container status \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": rpc error: code = NotFound desc = could not find container \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": container with ID starting with 46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.794823 4992 scope.go:117] "RemoveContainer" containerID="f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.795183 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} err="failed to get container status \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": rpc error: code = NotFound desc = could not find container \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": container with ID starting with f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.795210 4992 scope.go:117] "RemoveContainer" containerID="24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.795499 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} err="failed to get container status \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": rpc error: code = NotFound desc = could not find container \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": container with ID starting with 24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.795520 4992 scope.go:117] "RemoveContainer" containerID="6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.795747 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} err="failed to get container status \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": rpc error: code = NotFound desc = could not find container \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": container with ID starting with 6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.795768 4992 scope.go:117] "RemoveContainer" containerID="9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.796022 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} err="failed to get container status \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": rpc error: code = NotFound desc = could not find container \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": container with ID starting with 9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.796046 4992 scope.go:117] "RemoveContainer" containerID="db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.796340 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} err="failed to get container status \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": rpc error: code = NotFound desc = could not find container \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": container with ID starting with db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.796387 4992 scope.go:117] "RemoveContainer" containerID="24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.796662 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} err="failed to get container status \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": rpc error: code = NotFound desc = could not find container \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": container with ID starting with 24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.796688 4992 scope.go:117] "RemoveContainer" containerID="93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.797067 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} err="failed to get container status \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": rpc error: code = NotFound desc = could not find container \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": container with ID starting with 93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.797107 4992 scope.go:117] "RemoveContainer" containerID="80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.797586 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} err="failed to get container status \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": rpc error: code = NotFound desc = could not find container \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": container with ID starting with 80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.797615 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.797865 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} err="failed to get container status \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": rpc error: code = NotFound desc = could not find container \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": container with ID starting with bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.797890 4992 scope.go:117] "RemoveContainer" containerID="46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.798307 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} err="failed to get container status \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": rpc error: code = NotFound desc = could not find container \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": container with ID starting with 46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.798337 4992 scope.go:117] "RemoveContainer" containerID="f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.798703 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} err="failed to get container status \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": rpc error: code = NotFound desc = could not find container \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": container with ID starting with f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.798849 4992 scope.go:117] "RemoveContainer" containerID="24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.799206 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} err="failed to get container status \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": rpc error: code = NotFound desc = could not find container \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": container with ID starting with 24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.799232 4992 scope.go:117] "RemoveContainer" containerID="6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.799530 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} err="failed to get container status \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": rpc error: code = NotFound desc = could not find container \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": container with ID starting with 6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.799555 4992 scope.go:117] "RemoveContainer" containerID="9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.799874 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} err="failed to get container status \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": rpc error: code = NotFound desc = could not find container \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": container with ID starting with 9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.799904 4992 scope.go:117] "RemoveContainer" containerID="db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.800232 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} err="failed to get container status \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": rpc error: code = NotFound desc = could not find container \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": container with ID starting with db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.800257 4992 scope.go:117] "RemoveContainer" containerID="24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.800571 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} err="failed to get container status \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": rpc error: code = NotFound desc = could not find container \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": container with ID starting with 24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.800606 4992 scope.go:117] "RemoveContainer" containerID="93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.800901 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} err="failed to get container status \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": rpc error: code = NotFound desc = could not find container \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": container with ID starting with 93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.800949 4992 scope.go:117] "RemoveContainer" containerID="80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.801418 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720"} err="failed to get container status \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": rpc error: code = NotFound desc = could not find container \"80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720\": container with ID starting with 80c8d570c7f3ab58d3b6c7f642b1a4f1d2cf6fa03c31a85331f3e0aa814fb720 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.801452 4992 scope.go:117] "RemoveContainer" containerID="bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.801757 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7"} err="failed to get container status \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": rpc error: code = NotFound desc = could not find container \"bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7\": container with ID starting with bf1e311c423febf086b4164a28b72c3fbfb5b6c4c89ae65641b8ec240521ede7 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.801788 4992 scope.go:117] "RemoveContainer" containerID="46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.802074 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9"} err="failed to get container status \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": rpc error: code = NotFound desc = could not find container \"46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9\": container with ID starting with 46cb3d61dbb9d51dafb4f12f7238cdde303ec3f240d643ea7ffee32a156a2dd9 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.802117 4992 scope.go:117] "RemoveContainer" containerID="f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.802964 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f"} err="failed to get container status \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": rpc error: code = NotFound desc = could not find container \"f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f\": container with ID starting with f6e8bd52ad16f0833ee8c4559ac9a231e545b079b7056158361f9f67f2307c1f not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.802989 4992 scope.go:117] "RemoveContainer" containerID="24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.803289 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f"} err="failed to get container status \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": rpc error: code = NotFound desc = could not find container \"24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f\": container with ID starting with 24f6732724bb07481ff2ab6551014e4581dfa8e73e5f2e65f1ec3c9a7e78935f not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.803334 4992 scope.go:117] "RemoveContainer" containerID="6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.803787 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc"} err="failed to get container status \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": rpc error: code = NotFound desc = could not find container \"6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc\": container with ID starting with 6dc4b5ff7b45112821fde7bad0426080b7a3e40fb084d8e7a9ef3ceca41ba4bc not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.803818 4992 scope.go:117] "RemoveContainer" containerID="9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.804124 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c"} err="failed to get container status \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": rpc error: code = NotFound desc = could not find container \"9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c\": container with ID starting with 9f5c18bfd1a22aa6a6ecc2c799625256548fc1527172a6e4e9574f3bc757799c not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.804150 4992 scope.go:117] "RemoveContainer" containerID="db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.804510 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c"} err="failed to get container status \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": rpc error: code = NotFound desc = could not find container \"db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c\": container with ID starting with db6d2348536073696c0b61c7d75a1adb407d7d7991a7dd6ab23beb9daa060e1c not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.804537 4992 scope.go:117] "RemoveContainer" containerID="24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.804832 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940"} err="failed to get container status \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": rpc error: code = NotFound desc = could not find container \"24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940\": container with ID starting with 24ef399aebc32f33f8058e049ef7fce8bd812d807ac04c4972f00817aa549940 not found: ID does not exist" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.804857 4992 scope.go:117] "RemoveContainer" containerID="93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3" Oct 04 04:46:30 crc kubenswrapper[4992]: I1004 04:46:30.805134 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3"} err="failed to get container status \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": rpc error: code = NotFound desc = could not find container \"93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3\": container with ID starting with 93b95cb4a5b8fe52853183657b1f06c0b0a225218cc340840171e458b43293e3 not found: ID does not exist" Oct 04 04:46:31 crc kubenswrapper[4992]: I1004 04:46:31.289590 4992 generic.go:334] "Generic (PLEG): container finished" podID="b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22" containerID="c9c4574fd53fe0dbc5dc0e1b2282691a4dcf83cb6a5788ebc3f59672d2fd7738" exitCode=0 Oct 04 04:46:31 crc kubenswrapper[4992]: I1004 04:46:31.289658 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerDied","Data":"c9c4574fd53fe0dbc5dc0e1b2282691a4dcf83cb6a5788ebc3f59672d2fd7738"} Oct 04 04:46:31 crc kubenswrapper[4992]: I1004 04:46:31.292819 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vhq64_2b6879c5-46f4-4d87-959b-03dff08cff4c/kube-multus/2.log" Oct 04 04:46:31 crc kubenswrapper[4992]: I1004 04:46:31.292881 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vhq64" event={"ID":"2b6879c5-46f4-4d87-959b-03dff08cff4c","Type":"ContainerStarted","Data":"8a2054ca1fd2df28fb5e62c4742bc4b525d87d2f1d9a8d946558955276c61ba7"} Oct 04 04:46:32 crc kubenswrapper[4992]: I1004 04:46:32.306435 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"37321237c58c820d026450d89135fd2601bc848eca205c0240004fb80d63ce2c"} Oct 04 04:46:32 crc kubenswrapper[4992]: I1004 04:46:32.306852 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"c0538017a0e2e852ef51287b1d079860a9ec100599f55f6036821dc5667bb73c"} Oct 04 04:46:32 crc kubenswrapper[4992]: I1004 04:46:32.306876 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"16075c2bd33c2ab3e1c5b45a16b15942a5dc7bf5bb6dc750cf676fc1d7d0c9a5"} Oct 04 04:46:32 crc kubenswrapper[4992]: I1004 04:46:32.306894 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"d2b5a501e245b06d378efbc48bcc8f1a5a1485589de61c44e6cae15dd0edce96"} Oct 04 04:46:32 crc kubenswrapper[4992]: I1004 04:46:32.306912 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"0e3c50218249987b96c2f71b547512c53420f0a08ba884c3e19eeb4534053e79"} Oct 04 04:46:32 crc kubenswrapper[4992]: I1004 04:46:32.306930 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"3d069287e53b497f492c77708e95bb20cec67547eec034f7999d3649d374967c"} Oct 04 04:46:32 crc kubenswrapper[4992]: I1004 04:46:32.331373 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e301ed8-9506-4d6e-bbce-d2ad3817ca36" path="/var/lib/kubelet/pods/2e301ed8-9506-4d6e-bbce-d2ad3817ca36/volumes" Oct 04 04:46:34 crc kubenswrapper[4992]: I1004 04:46:34.329568 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"679f50b932707cd926fcd56bd37115c4b69b7217e005615a31694e0a3c00508b"} Oct 04 04:46:37 crc kubenswrapper[4992]: I1004 04:46:37.344793 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" event={"ID":"b59d8f64-44d4-4dfb-b6e3-bdcca8ecce22","Type":"ContainerStarted","Data":"fb6aeabd6d46d70f0ae810703fecf38a20c02d9d44161123262c5fe97d49a9f7"} Oct 04 04:46:37 crc kubenswrapper[4992]: I1004 04:46:37.346068 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:37 crc kubenswrapper[4992]: I1004 04:46:37.346103 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:37 crc kubenswrapper[4992]: I1004 04:46:37.346136 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:37 crc kubenswrapper[4992]: I1004 04:46:37.371961 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:37 crc kubenswrapper[4992]: I1004 04:46:37.373064 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:46:37 crc kubenswrapper[4992]: I1004 04:46:37.416767 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" podStartSLOduration=8.416749546 podStartE2EDuration="8.416749546s" podCreationTimestamp="2025-10-04 04:46:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:46:37.3812201 +0000 UTC m=+831.228895568" watchObservedRunningTime="2025-10-04 04:46:37.416749546 +0000 UTC m=+831.264425024" Oct 04 04:46:50 crc kubenswrapper[4992]: I1004 04:46:50.894894 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:46:50 crc kubenswrapper[4992]: I1004 04:46:50.895674 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:47:00 crc kubenswrapper[4992]: I1004 04:47:00.232881 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pg9kr" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.132893 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn"] Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.134497 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.136951 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.146522 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn"] Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.248325 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg25p\" (UniqueName: \"kubernetes.io/projected/a5658ee6-1e89-477f-a4d3-f871bd8ca728-kube-api-access-mg25p\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.248430 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.248459 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.350245 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg25p\" (UniqueName: \"kubernetes.io/projected/a5658ee6-1e89-477f-a4d3-f871bd8ca728-kube-api-access-mg25p\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.350304 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.350337 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.350703 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.350762 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.367464 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg25p\" (UniqueName: \"kubernetes.io/projected/a5658ee6-1e89-477f-a4d3-f871bd8ca728-kube-api-access-mg25p\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.468231 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:08 crc kubenswrapper[4992]: I1004 04:47:08.905202 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn"] Oct 04 04:47:09 crc kubenswrapper[4992]: I1004 04:47:09.517847 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" event={"ID":"a5658ee6-1e89-477f-a4d3-f871bd8ca728","Type":"ContainerStarted","Data":"a293b0a005302770fb3bb5c0377bbd39bb3b1f8905583081a27fd347c8ae77d4"} Oct 04 04:47:09 crc kubenswrapper[4992]: I1004 04:47:09.518251 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" event={"ID":"a5658ee6-1e89-477f-a4d3-f871bd8ca728","Type":"ContainerStarted","Data":"99bd140d053d55f5f78b7042d36682a693f52bb907a0d01b44fc42ff176221e6"} Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.508032 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bccln"] Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.509495 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.514840 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bccln"] Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.526273 4992 generic.go:334] "Generic (PLEG): container finished" podID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerID="a293b0a005302770fb3bb5c0377bbd39bb3b1f8905583081a27fd347c8ae77d4" exitCode=0 Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.526477 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" event={"ID":"a5658ee6-1e89-477f-a4d3-f871bd8ca728","Type":"ContainerDied","Data":"a293b0a005302770fb3bb5c0377bbd39bb3b1f8905583081a27fd347c8ae77d4"} Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.681845 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-catalog-content\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.681906 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-utilities\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.681940 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnw6f\" (UniqueName: \"kubernetes.io/projected/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-kube-api-access-xnw6f\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.783545 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-catalog-content\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.783644 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-utilities\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.783683 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnw6f\" (UniqueName: \"kubernetes.io/projected/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-kube-api-access-xnw6f\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.784123 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-catalog-content\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.784424 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-utilities\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.805153 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnw6f\" (UniqueName: \"kubernetes.io/projected/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-kube-api-access-xnw6f\") pod \"redhat-operators-bccln\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:10 crc kubenswrapper[4992]: I1004 04:47:10.836963 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:11 crc kubenswrapper[4992]: I1004 04:47:11.282374 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bccln"] Oct 04 04:47:11 crc kubenswrapper[4992]: I1004 04:47:11.533703 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bccln" event={"ID":"fd9f9935-fd71-43cd-a0f2-4bb8751cad59","Type":"ContainerStarted","Data":"9a9202967bc14c689d53c95c4a72962445edaad735402efc2e221d205030c539"} Oct 04 04:47:12 crc kubenswrapper[4992]: I1004 04:47:12.543317 4992 generic.go:334] "Generic (PLEG): container finished" podID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerID="596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc" exitCode=0 Oct 04 04:47:12 crc kubenswrapper[4992]: I1004 04:47:12.543442 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bccln" event={"ID":"fd9f9935-fd71-43cd-a0f2-4bb8751cad59","Type":"ContainerDied","Data":"596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc"} Oct 04 04:47:12 crc kubenswrapper[4992]: I1004 04:47:12.548128 4992 generic.go:334] "Generic (PLEG): container finished" podID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerID="0036af44fc983793a47feda9562dd38c8d33b5350f38766258f8507d4d0c25b2" exitCode=0 Oct 04 04:47:12 crc kubenswrapper[4992]: I1004 04:47:12.548204 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" event={"ID":"a5658ee6-1e89-477f-a4d3-f871bd8ca728","Type":"ContainerDied","Data":"0036af44fc983793a47feda9562dd38c8d33b5350f38766258f8507d4d0c25b2"} Oct 04 04:47:13 crc kubenswrapper[4992]: I1004 04:47:13.557731 4992 generic.go:334] "Generic (PLEG): container finished" podID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerID="5b1630afe7723c0e5d8f9cffa25ad37a315bb3ba29c8eab58d2c7c0321f087cd" exitCode=0 Oct 04 04:47:13 crc kubenswrapper[4992]: I1004 04:47:13.557774 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" event={"ID":"a5658ee6-1e89-477f-a4d3-f871bd8ca728","Type":"ContainerDied","Data":"5b1630afe7723c0e5d8f9cffa25ad37a315bb3ba29c8eab58d2c7c0321f087cd"} Oct 04 04:47:13 crc kubenswrapper[4992]: I1004 04:47:13.561010 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bccln" event={"ID":"fd9f9935-fd71-43cd-a0f2-4bb8751cad59","Type":"ContainerStarted","Data":"f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779"} Oct 04 04:47:14 crc kubenswrapper[4992]: E1004 04:47:14.075773 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd9f9935_fd71_43cd_a0f2_4bb8751cad59.slice/crio-f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd9f9935_fd71_43cd_a0f2_4bb8751cad59.slice/crio-conmon-f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:47:14 crc kubenswrapper[4992]: I1004 04:47:14.571042 4992 generic.go:334] "Generic (PLEG): container finished" podID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerID="f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779" exitCode=0 Oct 04 04:47:14 crc kubenswrapper[4992]: I1004 04:47:14.571147 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bccln" event={"ID":"fd9f9935-fd71-43cd-a0f2-4bb8751cad59","Type":"ContainerDied","Data":"f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779"} Oct 04 04:47:14 crc kubenswrapper[4992]: I1004 04:47:14.931601 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.043685 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg25p\" (UniqueName: \"kubernetes.io/projected/a5658ee6-1e89-477f-a4d3-f871bd8ca728-kube-api-access-mg25p\") pod \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.043742 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-util\") pod \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.043809 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-bundle\") pod \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\" (UID: \"a5658ee6-1e89-477f-a4d3-f871bd8ca728\") " Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.044460 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-bundle" (OuterVolumeSpecName: "bundle") pod "a5658ee6-1e89-477f-a4d3-f871bd8ca728" (UID: "a5658ee6-1e89-477f-a4d3-f871bd8ca728"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.049505 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5658ee6-1e89-477f-a4d3-f871bd8ca728-kube-api-access-mg25p" (OuterVolumeSpecName: "kube-api-access-mg25p") pod "a5658ee6-1e89-477f-a4d3-f871bd8ca728" (UID: "a5658ee6-1e89-477f-a4d3-f871bd8ca728"). InnerVolumeSpecName "kube-api-access-mg25p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.100121 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-util" (OuterVolumeSpecName: "util") pod "a5658ee6-1e89-477f-a4d3-f871bd8ca728" (UID: "a5658ee6-1e89-477f-a4d3-f871bd8ca728"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.145105 4992 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.145134 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg25p\" (UniqueName: \"kubernetes.io/projected/a5658ee6-1e89-477f-a4d3-f871bd8ca728-kube-api-access-mg25p\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.145147 4992 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5658ee6-1e89-477f-a4d3-f871bd8ca728-util\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.581524 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" event={"ID":"a5658ee6-1e89-477f-a4d3-f871bd8ca728","Type":"ContainerDied","Data":"99bd140d053d55f5f78b7042d36682a693f52bb907a0d01b44fc42ff176221e6"} Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.581542 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.581769 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99bd140d053d55f5f78b7042d36682a693f52bb907a0d01b44fc42ff176221e6" Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.583446 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bccln" event={"ID":"fd9f9935-fd71-43cd-a0f2-4bb8751cad59","Type":"ContainerStarted","Data":"5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac"} Oct 04 04:47:15 crc kubenswrapper[4992]: I1004 04:47:15.615423 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bccln" podStartSLOduration=3.026235898 podStartE2EDuration="5.615345176s" podCreationTimestamp="2025-10-04 04:47:10 +0000 UTC" firstStartedPulling="2025-10-04 04:47:12.545176103 +0000 UTC m=+866.392851601" lastFinishedPulling="2025-10-04 04:47:15.134285391 +0000 UTC m=+868.981960879" observedRunningTime="2025-10-04 04:47:15.606344666 +0000 UTC m=+869.454020164" watchObservedRunningTime="2025-10-04 04:47:15.615345176 +0000 UTC m=+869.463020684" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.776233 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6"] Oct 04 04:47:19 crc kubenswrapper[4992]: E1004 04:47:19.777025 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerName="pull" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.777041 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerName="pull" Oct 04 04:47:19 crc kubenswrapper[4992]: E1004 04:47:19.777059 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerName="util" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.777067 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerName="util" Oct 04 04:47:19 crc kubenswrapper[4992]: E1004 04:47:19.777084 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerName="extract" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.777093 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerName="extract" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.777208 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5658ee6-1e89-477f-a4d3-f871bd8ca728" containerName="extract" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.777665 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.779089 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-8sgzr" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.779524 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.779590 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.801468 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6"] Oct 04 04:47:19 crc kubenswrapper[4992]: I1004 04:47:19.914610 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmwc2\" (UniqueName: \"kubernetes.io/projected/dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a-kube-api-access-qmwc2\") pod \"nmstate-operator-858ddd8f98-kr2c6\" (UID: \"dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.015449 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmwc2\" (UniqueName: \"kubernetes.io/projected/dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a-kube-api-access-qmwc2\") pod \"nmstate-operator-858ddd8f98-kr2c6\" (UID: \"dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.033724 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmwc2\" (UniqueName: \"kubernetes.io/projected/dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a-kube-api-access-qmwc2\") pod \"nmstate-operator-858ddd8f98-kr2c6\" (UID: \"dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.093219 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.514117 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6"] Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.609591 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6" event={"ID":"dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a","Type":"ContainerStarted","Data":"0fac966b8ad1ae842bcc8cb8158958174024a5d7eea6ed6202b38e7a4db2ea24"} Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.838005 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.838288 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.888105 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.894584 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.894650 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.894697 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.895335 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"391734b7432560d7fffa75f76956def8a70e770707e3c21215c4ea365ca34204"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:47:20 crc kubenswrapper[4992]: I1004 04:47:20.895423 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://391734b7432560d7fffa75f76956def8a70e770707e3c21215c4ea365ca34204" gracePeriod=600 Oct 04 04:47:21 crc kubenswrapper[4992]: I1004 04:47:21.659784 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:22 crc kubenswrapper[4992]: I1004 04:47:22.622566 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="391734b7432560d7fffa75f76956def8a70e770707e3c21215c4ea365ca34204" exitCode=0 Oct 04 04:47:22 crc kubenswrapper[4992]: I1004 04:47:22.622645 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"391734b7432560d7fffa75f76956def8a70e770707e3c21215c4ea365ca34204"} Oct 04 04:47:22 crc kubenswrapper[4992]: I1004 04:47:22.622700 4992 scope.go:117] "RemoveContainer" containerID="34bbae1fd997a7077d97cc7f2a26b45479ea4683901678b60f3f46c1a59664b1" Oct 04 04:47:23 crc kubenswrapper[4992]: I1004 04:47:23.295427 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bccln"] Oct 04 04:47:23 crc kubenswrapper[4992]: I1004 04:47:23.630822 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"bb88432e3860828bfc70223a5d4820bc8d7427ffafdb37bf5816fc3fffccf08d"} Oct 04 04:47:24 crc kubenswrapper[4992]: I1004 04:47:24.635659 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6" event={"ID":"dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a","Type":"ContainerStarted","Data":"767638b53d958b7d1e6ee741ba2de2669f496d8a55610d60eec957031aaaf2e3"} Oct 04 04:47:24 crc kubenswrapper[4992]: I1004 04:47:24.635820 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bccln" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerName="registry-server" containerID="cri-o://5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac" gracePeriod=2 Oct 04 04:47:24 crc kubenswrapper[4992]: I1004 04:47:24.657866 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-kr2c6" podStartSLOduration=2.516667966 podStartE2EDuration="5.657843948s" podCreationTimestamp="2025-10-04 04:47:19 +0000 UTC" firstStartedPulling="2025-10-04 04:47:20.52538407 +0000 UTC m=+874.373059538" lastFinishedPulling="2025-10-04 04:47:23.666560052 +0000 UTC m=+877.514235520" observedRunningTime="2025-10-04 04:47:24.655076454 +0000 UTC m=+878.502751942" watchObservedRunningTime="2025-10-04 04:47:24.657843948 +0000 UTC m=+878.505519426" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.053669 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.187017 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnw6f\" (UniqueName: \"kubernetes.io/projected/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-kube-api-access-xnw6f\") pod \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.187109 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-catalog-content\") pod \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.187189 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-utilities\") pod \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\" (UID: \"fd9f9935-fd71-43cd-a0f2-4bb8751cad59\") " Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.188068 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-utilities" (OuterVolumeSpecName: "utilities") pod "fd9f9935-fd71-43cd-a0f2-4bb8751cad59" (UID: "fd9f9935-fd71-43cd-a0f2-4bb8751cad59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.200345 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-kube-api-access-xnw6f" (OuterVolumeSpecName: "kube-api-access-xnw6f") pod "fd9f9935-fd71-43cd-a0f2-4bb8751cad59" (UID: "fd9f9935-fd71-43cd-a0f2-4bb8751cad59"). InnerVolumeSpecName "kube-api-access-xnw6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.271276 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd9f9935-fd71-43cd-a0f2-4bb8751cad59" (UID: "fd9f9935-fd71-43cd-a0f2-4bb8751cad59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.288228 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.288268 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnw6f\" (UniqueName: \"kubernetes.io/projected/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-kube-api-access-xnw6f\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.288279 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd9f9935-fd71-43cd-a0f2-4bb8751cad59-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.642826 4992 generic.go:334] "Generic (PLEG): container finished" podID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerID="5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac" exitCode=0 Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.642905 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bccln" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.642919 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bccln" event={"ID":"fd9f9935-fd71-43cd-a0f2-4bb8751cad59","Type":"ContainerDied","Data":"5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac"} Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.643209 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bccln" event={"ID":"fd9f9935-fd71-43cd-a0f2-4bb8751cad59","Type":"ContainerDied","Data":"9a9202967bc14c689d53c95c4a72962445edaad735402efc2e221d205030c539"} Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.643239 4992 scope.go:117] "RemoveContainer" containerID="5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.664028 4992 scope.go:117] "RemoveContainer" containerID="f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.668411 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bccln"] Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.674076 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bccln"] Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.700970 4992 scope.go:117] "RemoveContainer" containerID="596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.713501 4992 scope.go:117] "RemoveContainer" containerID="5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac" Oct 04 04:47:25 crc kubenswrapper[4992]: E1004 04:47:25.713884 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac\": container with ID starting with 5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac not found: ID does not exist" containerID="5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.713911 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac"} err="failed to get container status \"5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac\": rpc error: code = NotFound desc = could not find container \"5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac\": container with ID starting with 5597aba9d742520b8e20e47632ffeef6fab8c90747a65f4bad629a8855a5bbac not found: ID does not exist" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.713930 4992 scope.go:117] "RemoveContainer" containerID="f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779" Oct 04 04:47:25 crc kubenswrapper[4992]: E1004 04:47:25.714254 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779\": container with ID starting with f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779 not found: ID does not exist" containerID="f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.714305 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779"} err="failed to get container status \"f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779\": rpc error: code = NotFound desc = could not find container \"f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779\": container with ID starting with f66307ebdbf27c2a0d22806bc41e85471f513423ac780d358ab7edfe43549779 not found: ID does not exist" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.714346 4992 scope.go:117] "RemoveContainer" containerID="596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc" Oct 04 04:47:25 crc kubenswrapper[4992]: E1004 04:47:25.714630 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc\": container with ID starting with 596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc not found: ID does not exist" containerID="596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc" Oct 04 04:47:25 crc kubenswrapper[4992]: I1004 04:47:25.714654 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc"} err="failed to get container status \"596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc\": rpc error: code = NotFound desc = could not find container \"596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc\": container with ID starting with 596b3fdb8b55eaf1d5e0d953030f97e2d703156fb9b4237af2c05e4303ee43cc not found: ID does not exist" Oct 04 04:47:26 crc kubenswrapper[4992]: I1004 04:47:26.334382 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" path="/var/lib/kubelet/pods/fd9f9935-fd71-43cd-a0f2-4bb8751cad59/volumes" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.581499 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9"] Oct 04 04:47:28 crc kubenswrapper[4992]: E1004 04:47:28.582023 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerName="extract-content" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.582036 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerName="extract-content" Oct 04 04:47:28 crc kubenswrapper[4992]: E1004 04:47:28.582048 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerName="extract-utilities" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.582053 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerName="extract-utilities" Oct 04 04:47:28 crc kubenswrapper[4992]: E1004 04:47:28.582063 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerName="registry-server" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.582070 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerName="registry-server" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.582205 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd9f9935-fd71-43cd-a0f2-4bb8751cad59" containerName="registry-server" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.583053 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.586032 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-4l9kt" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.589978 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9"] Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.604838 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z"] Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.605867 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.607259 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.617198 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z"] Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.622618 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-xn8rv"] Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.623242 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.711939 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp"] Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.713098 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.714886 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.715998 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.716205 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-njcpf" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.720687 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp"] Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.729558 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/00d2e67f-7b6a-43dd-afd0-500f16e5ae3c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-kvc8z\" (UID: \"00d2e67f-7b6a-43dd-afd0-500f16e5ae3c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.729605 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vtpw\" (UniqueName: \"kubernetes.io/projected/00d2e67f-7b6a-43dd-afd0-500f16e5ae3c-kube-api-access-8vtpw\") pod \"nmstate-webhook-6cdbc54649-kvc8z\" (UID: \"00d2e67f-7b6a-43dd-afd0-500f16e5ae3c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.729636 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdlmb\" (UniqueName: \"kubernetes.io/projected/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-kube-api-access-fdlmb\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.729671 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-nmstate-lock\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.729691 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-dbus-socket\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.729725 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-ovs-socket\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.729761 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6gbd\" (UniqueName: \"kubernetes.io/projected/c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0-kube-api-access-j6gbd\") pod \"nmstate-metrics-fdff9cb8d-qncn9\" (UID: \"c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831292 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdlmb\" (UniqueName: \"kubernetes.io/projected/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-kube-api-access-fdlmb\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831434 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3bba2412-53af-4a50-aa28-439c21630dd6-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831496 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-nmstate-lock\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831525 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-dbus-socket\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831553 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3bba2412-53af-4a50-aa28-439c21630dd6-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831584 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-ovs-socket\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831638 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6gbd\" (UniqueName: \"kubernetes.io/projected/c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0-kube-api-access-j6gbd\") pod \"nmstate-metrics-fdff9cb8d-qncn9\" (UID: \"c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831653 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-nmstate-lock\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.831768 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/00d2e67f-7b6a-43dd-afd0-500f16e5ae3c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-kvc8z\" (UID: \"00d2e67f-7b6a-43dd-afd0-500f16e5ae3c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.832095 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-dbus-socket\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.832131 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-ovs-socket\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.832167 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftfvm\" (UniqueName: \"kubernetes.io/projected/3bba2412-53af-4a50-aa28-439c21630dd6-kube-api-access-ftfvm\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.832298 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vtpw\" (UniqueName: \"kubernetes.io/projected/00d2e67f-7b6a-43dd-afd0-500f16e5ae3c-kube-api-access-8vtpw\") pod \"nmstate-webhook-6cdbc54649-kvc8z\" (UID: \"00d2e67f-7b6a-43dd-afd0-500f16e5ae3c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.839076 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/00d2e67f-7b6a-43dd-afd0-500f16e5ae3c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-kvc8z\" (UID: \"00d2e67f-7b6a-43dd-afd0-500f16e5ae3c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.854985 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdlmb\" (UniqueName: \"kubernetes.io/projected/e65ecf64-2e05-4ce7-a67a-ec85b8c39d67-kube-api-access-fdlmb\") pod \"nmstate-handler-xn8rv\" (UID: \"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67\") " pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.866213 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6gbd\" (UniqueName: \"kubernetes.io/projected/c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0-kube-api-access-j6gbd\") pod \"nmstate-metrics-fdff9cb8d-qncn9\" (UID: \"c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.868726 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vtpw\" (UniqueName: \"kubernetes.io/projected/00d2e67f-7b6a-43dd-afd0-500f16e5ae3c-kube-api-access-8vtpw\") pod \"nmstate-webhook-6cdbc54649-kvc8z\" (UID: \"00d2e67f-7b6a-43dd-afd0-500f16e5ae3c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.899852 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.906325 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5755ff94dd-mlmlj"] Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.907201 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.919631 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.936395 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3bba2412-53af-4a50-aa28-439c21630dd6-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.936449 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3bba2412-53af-4a50-aa28-439c21630dd6-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.936508 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftfvm\" (UniqueName: \"kubernetes.io/projected/3bba2412-53af-4a50-aa28-439c21630dd6-kube-api-access-ftfvm\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.937673 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3bba2412-53af-4a50-aa28-439c21630dd6-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.939979 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.942449 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3bba2412-53af-4a50-aa28-439c21630dd6-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.961564 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftfvm\" (UniqueName: \"kubernetes.io/projected/3bba2412-53af-4a50-aa28-439c21630dd6-kube-api-access-ftfvm\") pod \"nmstate-console-plugin-6b874cbd85-s2crp\" (UID: \"3bba2412-53af-4a50-aa28-439c21630dd6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:28 crc kubenswrapper[4992]: W1004 04:47:28.967074 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode65ecf64_2e05_4ce7_a67a_ec85b8c39d67.slice/crio-a51c77a2605038ca3e259065e6d24f4b00b08e286f357f2cfa4803d8274affe7 WatchSource:0}: Error finding container a51c77a2605038ca3e259065e6d24f4b00b08e286f357f2cfa4803d8274affe7: Status 404 returned error can't find the container with id a51c77a2605038ca3e259065e6d24f4b00b08e286f357f2cfa4803d8274affe7 Oct 04 04:47:28 crc kubenswrapper[4992]: I1004 04:47:28.981239 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5755ff94dd-mlmlj"] Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.029457 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.040343 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmpxx\" (UniqueName: \"kubernetes.io/projected/f7dfc802-6699-4857-b37a-998d3ce7f69a-kube-api-access-bmpxx\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.040401 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-oauth-serving-cert\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.040429 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-service-ca\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.040446 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-trusted-ca-bundle\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.040498 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-serving-cert\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.040524 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-config\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.040572 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-oauth-config\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.142263 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-oauth-serving-cert\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.142309 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-service-ca\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.142334 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-trusted-ca-bundle\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.142396 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-serving-cert\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.142418 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-config\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.142442 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-oauth-config\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.142467 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmpxx\" (UniqueName: \"kubernetes.io/projected/f7dfc802-6699-4857-b37a-998d3ce7f69a-kube-api-access-bmpxx\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.145224 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-config\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.145753 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-oauth-serving-cert\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.145803 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-service-ca\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.146054 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7dfc802-6699-4857-b37a-998d3ce7f69a-trusted-ca-bundle\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.157766 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-oauth-config\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.158322 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7dfc802-6699-4857-b37a-998d3ce7f69a-console-serving-cert\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.165978 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmpxx\" (UniqueName: \"kubernetes.io/projected/f7dfc802-6699-4857-b37a-998d3ce7f69a-kube-api-access-bmpxx\") pod \"console-5755ff94dd-mlmlj\" (UID: \"f7dfc802-6699-4857-b37a-998d3ce7f69a\") " pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.272591 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.343348 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9"] Oct 04 04:47:29 crc kubenswrapper[4992]: W1004 04:47:29.358635 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1f8f275_8e3a_4ed4_ae06_7a14cd5b93f0.slice/crio-ba08432721fbdba3758f2db123e9aee7fce660d2b17a2ad63a936a0a248d8a7e WatchSource:0}: Error finding container ba08432721fbdba3758f2db123e9aee7fce660d2b17a2ad63a936a0a248d8a7e: Status 404 returned error can't find the container with id ba08432721fbdba3758f2db123e9aee7fce660d2b17a2ad63a936a0a248d8a7e Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.406224 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z"] Oct 04 04:47:29 crc kubenswrapper[4992]: W1004 04:47:29.413724 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00d2e67f_7b6a_43dd_afd0_500f16e5ae3c.slice/crio-f7f752e26b1270a466f933f397935bc17beb1c194f00f371a5d5edbc5a2eb09c WatchSource:0}: Error finding container f7f752e26b1270a466f933f397935bc17beb1c194f00f371a5d5edbc5a2eb09c: Status 404 returned error can't find the container with id f7f752e26b1270a466f933f397935bc17beb1c194f00f371a5d5edbc5a2eb09c Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.470660 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp"] Oct 04 04:47:29 crc kubenswrapper[4992]: W1004 04:47:29.474072 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bba2412_53af_4a50_aa28_439c21630dd6.slice/crio-e6099ea089f7cfc55c0b10fc339db4b0a64531c7c641e4d9a7a21fb58c85d262 WatchSource:0}: Error finding container e6099ea089f7cfc55c0b10fc339db4b0a64531c7c641e4d9a7a21fb58c85d262: Status 404 returned error can't find the container with id e6099ea089f7cfc55c0b10fc339db4b0a64531c7c641e4d9a7a21fb58c85d262 Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.659881 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5755ff94dd-mlmlj"] Oct 04 04:47:29 crc kubenswrapper[4992]: W1004 04:47:29.664814 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7dfc802_6699_4857_b37a_998d3ce7f69a.slice/crio-143683be4cb730555c5bba771e2c201b400c1f9ed8a8b47b3f6159a1d99ff7cb WatchSource:0}: Error finding container 143683be4cb730555c5bba771e2c201b400c1f9ed8a8b47b3f6159a1d99ff7cb: Status 404 returned error can't find the container with id 143683be4cb730555c5bba771e2c201b400c1f9ed8a8b47b3f6159a1d99ff7cb Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.667816 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" event={"ID":"3bba2412-53af-4a50-aa28-439c21630dd6","Type":"ContainerStarted","Data":"e6099ea089f7cfc55c0b10fc339db4b0a64531c7c641e4d9a7a21fb58c85d262"} Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.672976 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" event={"ID":"00d2e67f-7b6a-43dd-afd0-500f16e5ae3c","Type":"ContainerStarted","Data":"f7f752e26b1270a466f933f397935bc17beb1c194f00f371a5d5edbc5a2eb09c"} Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.674232 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xn8rv" event={"ID":"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67","Type":"ContainerStarted","Data":"a51c77a2605038ca3e259065e6d24f4b00b08e286f357f2cfa4803d8274affe7"} Oct 04 04:47:29 crc kubenswrapper[4992]: I1004 04:47:29.675141 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" event={"ID":"c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0","Type":"ContainerStarted","Data":"ba08432721fbdba3758f2db123e9aee7fce660d2b17a2ad63a936a0a248d8a7e"} Oct 04 04:47:30 crc kubenswrapper[4992]: I1004 04:47:30.686574 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5755ff94dd-mlmlj" event={"ID":"f7dfc802-6699-4857-b37a-998d3ce7f69a","Type":"ContainerStarted","Data":"1cc0a2482730fb7b536e29585accb824f17f864944e5fa234fd1056a88db9397"} Oct 04 04:47:30 crc kubenswrapper[4992]: I1004 04:47:30.686993 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5755ff94dd-mlmlj" event={"ID":"f7dfc802-6699-4857-b37a-998d3ce7f69a","Type":"ContainerStarted","Data":"143683be4cb730555c5bba771e2c201b400c1f9ed8a8b47b3f6159a1d99ff7cb"} Oct 04 04:47:30 crc kubenswrapper[4992]: I1004 04:47:30.705728 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5755ff94dd-mlmlj" podStartSLOduration=2.705687758 podStartE2EDuration="2.705687758s" podCreationTimestamp="2025-10-04 04:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:47:30.705290848 +0000 UTC m=+884.552966366" watchObservedRunningTime="2025-10-04 04:47:30.705687758 +0000 UTC m=+884.553363226" Oct 04 04:47:31 crc kubenswrapper[4992]: I1004 04:47:31.695623 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-xn8rv" event={"ID":"e65ecf64-2e05-4ce7-a67a-ec85b8c39d67","Type":"ContainerStarted","Data":"82070e1d673ebc11697a3bc6eff74b441dcc6a37bb6a66950760734ae3d18794"} Oct 04 04:47:31 crc kubenswrapper[4992]: I1004 04:47:31.696296 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:31 crc kubenswrapper[4992]: I1004 04:47:31.698392 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" event={"ID":"c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0","Type":"ContainerStarted","Data":"7d556e3af4b1316287317e1c0bac2592f321df3e6fa6fe1f0c381923dda2e8a3"} Oct 04 04:47:31 crc kubenswrapper[4992]: I1004 04:47:31.700252 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" event={"ID":"00d2e67f-7b6a-43dd-afd0-500f16e5ae3c","Type":"ContainerStarted","Data":"7f569c60420de2dd77d463b997886aa1e3a6531029ebc31fdb55f77dfc3476a9"} Oct 04 04:47:31 crc kubenswrapper[4992]: I1004 04:47:31.727830 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-xn8rv" podStartSLOduration=1.380701101 podStartE2EDuration="3.727815355s" podCreationTimestamp="2025-10-04 04:47:28 +0000 UTC" firstStartedPulling="2025-10-04 04:47:28.976163553 +0000 UTC m=+882.823839021" lastFinishedPulling="2025-10-04 04:47:31.323277807 +0000 UTC m=+885.170953275" observedRunningTime="2025-10-04 04:47:31.719534035 +0000 UTC m=+885.567209503" watchObservedRunningTime="2025-10-04 04:47:31.727815355 +0000 UTC m=+885.575490823" Oct 04 04:47:31 crc kubenswrapper[4992]: I1004 04:47:31.736541 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" podStartSLOduration=1.8109483530000001 podStartE2EDuration="3.736520877s" podCreationTimestamp="2025-10-04 04:47:28 +0000 UTC" firstStartedPulling="2025-10-04 04:47:29.416408731 +0000 UTC m=+883.264084199" lastFinishedPulling="2025-10-04 04:47:31.341981255 +0000 UTC m=+885.189656723" observedRunningTime="2025-10-04 04:47:31.73588991 +0000 UTC m=+885.583565378" watchObservedRunningTime="2025-10-04 04:47:31.736520877 +0000 UTC m=+885.584196355" Oct 04 04:47:32 crc kubenswrapper[4992]: I1004 04:47:32.711802 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" event={"ID":"3bba2412-53af-4a50-aa28-439c21630dd6","Type":"ContainerStarted","Data":"e296798ad9a25047c7d320e13f54e6ebd3c10272453f7626b31ebf109dcf1b51"} Oct 04 04:47:32 crc kubenswrapper[4992]: I1004 04:47:32.712850 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:47:32 crc kubenswrapper[4992]: I1004 04:47:32.746002 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-s2crp" podStartSLOduration=1.955564042 podStartE2EDuration="4.745977527s" podCreationTimestamp="2025-10-04 04:47:28 +0000 UTC" firstStartedPulling="2025-10-04 04:47:29.476142341 +0000 UTC m=+883.323817809" lastFinishedPulling="2025-10-04 04:47:32.266555836 +0000 UTC m=+886.114231294" observedRunningTime="2025-10-04 04:47:32.735092877 +0000 UTC m=+886.582768345" watchObservedRunningTime="2025-10-04 04:47:32.745977527 +0000 UTC m=+886.593652995" Oct 04 04:47:34 crc kubenswrapper[4992]: I1004 04:47:34.727736 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" event={"ID":"c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0","Type":"ContainerStarted","Data":"64f6c0d0f32bb6b7166d992a3edbbb9f35cf688b69c79e07f8bc562a2232a6d3"} Oct 04 04:47:34 crc kubenswrapper[4992]: I1004 04:47:34.744907 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-qncn9" podStartSLOduration=2.371088091 podStartE2EDuration="6.744883633s" podCreationTimestamp="2025-10-04 04:47:28 +0000 UTC" firstStartedPulling="2025-10-04 04:47:29.361262773 +0000 UTC m=+883.208938241" lastFinishedPulling="2025-10-04 04:47:33.735058315 +0000 UTC m=+887.582733783" observedRunningTime="2025-10-04 04:47:34.744666457 +0000 UTC m=+888.592341975" watchObservedRunningTime="2025-10-04 04:47:34.744883633 +0000 UTC m=+888.592559141" Oct 04 04:47:38 crc kubenswrapper[4992]: I1004 04:47:38.965642 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-xn8rv" Oct 04 04:47:39 crc kubenswrapper[4992]: I1004 04:47:39.272994 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:39 crc kubenswrapper[4992]: I1004 04:47:39.273042 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:39 crc kubenswrapper[4992]: I1004 04:47:39.278970 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:39 crc kubenswrapper[4992]: I1004 04:47:39.765917 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5755ff94dd-mlmlj" Oct 04 04:47:39 crc kubenswrapper[4992]: I1004 04:47:39.843270 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5sgbq"] Oct 04 04:47:48 crc kubenswrapper[4992]: I1004 04:47:48.929617 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-kvc8z" Oct 04 04:48:04 crc kubenswrapper[4992]: I1004 04:48:04.916005 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5sgbq" podUID="8e339a50-fbf4-44da-9335-ba6ae033fc38" containerName="console" containerID="cri-o://19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b" gracePeriod=15 Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.299948 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5sgbq_8e339a50-fbf4-44da-9335-ba6ae033fc38/console/0.log" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.300302 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.386204 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjmg9\" (UniqueName: \"kubernetes.io/projected/8e339a50-fbf4-44da-9335-ba6ae033fc38-kube-api-access-hjmg9\") pod \"8e339a50-fbf4-44da-9335-ba6ae033fc38\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.386322 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-trusted-ca-bundle\") pod \"8e339a50-fbf4-44da-9335-ba6ae033fc38\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.386353 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-oauth-serving-cert\") pod \"8e339a50-fbf4-44da-9335-ba6ae033fc38\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.386440 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-oauth-config\") pod \"8e339a50-fbf4-44da-9335-ba6ae033fc38\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.386524 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-serving-cert\") pod \"8e339a50-fbf4-44da-9335-ba6ae033fc38\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.386545 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-config\") pod \"8e339a50-fbf4-44da-9335-ba6ae033fc38\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.386567 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-service-ca\") pod \"8e339a50-fbf4-44da-9335-ba6ae033fc38\" (UID: \"8e339a50-fbf4-44da-9335-ba6ae033fc38\") " Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.387959 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-service-ca" (OuterVolumeSpecName: "service-ca") pod "8e339a50-fbf4-44da-9335-ba6ae033fc38" (UID: "8e339a50-fbf4-44da-9335-ba6ae033fc38"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.387974 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-config" (OuterVolumeSpecName: "console-config") pod "8e339a50-fbf4-44da-9335-ba6ae033fc38" (UID: "8e339a50-fbf4-44da-9335-ba6ae033fc38"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.388026 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8e339a50-fbf4-44da-9335-ba6ae033fc38" (UID: "8e339a50-fbf4-44da-9335-ba6ae033fc38"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.388454 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8e339a50-fbf4-44da-9335-ba6ae033fc38" (UID: "8e339a50-fbf4-44da-9335-ba6ae033fc38"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.391784 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8e339a50-fbf4-44da-9335-ba6ae033fc38" (UID: "8e339a50-fbf4-44da-9335-ba6ae033fc38"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.391926 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8e339a50-fbf4-44da-9335-ba6ae033fc38" (UID: "8e339a50-fbf4-44da-9335-ba6ae033fc38"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.392287 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e339a50-fbf4-44da-9335-ba6ae033fc38-kube-api-access-hjmg9" (OuterVolumeSpecName: "kube-api-access-hjmg9") pod "8e339a50-fbf4-44da-9335-ba6ae033fc38" (UID: "8e339a50-fbf4-44da-9335-ba6ae033fc38"). InnerVolumeSpecName "kube-api-access-hjmg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.488190 4992 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.488249 4992 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.488270 4992 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.488290 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjmg9\" (UniqueName: \"kubernetes.io/projected/8e339a50-fbf4-44da-9335-ba6ae033fc38-kube-api-access-hjmg9\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.488311 4992 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.488332 4992 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e339a50-fbf4-44da-9335-ba6ae033fc38-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.488351 4992 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e339a50-fbf4-44da-9335-ba6ae033fc38-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.938533 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5sgbq_8e339a50-fbf4-44da-9335-ba6ae033fc38/console/0.log" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.938649 4992 generic.go:334] "Generic (PLEG): container finished" podID="8e339a50-fbf4-44da-9335-ba6ae033fc38" containerID="19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b" exitCode=2 Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.938707 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5sgbq" event={"ID":"8e339a50-fbf4-44da-9335-ba6ae033fc38","Type":"ContainerDied","Data":"19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b"} Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.938766 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5sgbq" event={"ID":"8e339a50-fbf4-44da-9335-ba6ae033fc38","Type":"ContainerDied","Data":"b24cf50aaf443a609f4671a6ceddb2b2af964f93c201e12ec69840dbe47ad7aa"} Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.938804 4992 scope.go:117] "RemoveContainer" containerID="19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.938810 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5sgbq" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.969293 4992 scope.go:117] "RemoveContainer" containerID="19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b" Oct 04 04:48:05 crc kubenswrapper[4992]: E1004 04:48:05.970033 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b\": container with ID starting with 19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b not found: ID does not exist" containerID="19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.970100 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b"} err="failed to get container status \"19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b\": rpc error: code = NotFound desc = could not find container \"19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b\": container with ID starting with 19f0042ba36f000f152d7b965e82ef06779e574ef14baf96f7ac747f947fc82b not found: ID does not exist" Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.989552 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5sgbq"] Oct 04 04:48:05 crc kubenswrapper[4992]: I1004 04:48:05.996199 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5sgbq"] Oct 04 04:48:06 crc kubenswrapper[4992]: I1004 04:48:06.333043 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e339a50-fbf4-44da-9335-ba6ae033fc38" path="/var/lib/kubelet/pods/8e339a50-fbf4-44da-9335-ba6ae033fc38/volumes" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.185613 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh"] Oct 04 04:48:07 crc kubenswrapper[4992]: E1004 04:48:07.185917 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e339a50-fbf4-44da-9335-ba6ae033fc38" containerName="console" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.185936 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e339a50-fbf4-44da-9335-ba6ae033fc38" containerName="console" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.186103 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e339a50-fbf4-44da-9335-ba6ae033fc38" containerName="console" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.187246 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.189165 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.203181 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh"] Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.312043 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.319935 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjwt9\" (UniqueName: \"kubernetes.io/projected/ca142652-0451-4c15-bd36-d2fc14a916ad-kube-api-access-kjwt9\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.320163 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.421179 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.421347 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.421459 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjwt9\" (UniqueName: \"kubernetes.io/projected/ca142652-0451-4c15-bd36-d2fc14a916ad-kube-api-access-kjwt9\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.422008 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.422076 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.444521 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjwt9\" (UniqueName: \"kubernetes.io/projected/ca142652-0451-4c15-bd36-d2fc14a916ad-kube-api-access-kjwt9\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.512758 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.929666 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh"] Oct 04 04:48:07 crc kubenswrapper[4992]: I1004 04:48:07.952208 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" event={"ID":"ca142652-0451-4c15-bd36-d2fc14a916ad","Type":"ContainerStarted","Data":"9b1fb3c24e3df03251c5c40bffe0c4f2431a1051d00d6ac6e356285f1b8e2e1d"} Oct 04 04:48:08 crc kubenswrapper[4992]: I1004 04:48:08.957940 4992 generic.go:334] "Generic (PLEG): container finished" podID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerID="8beaa871fe6518c9635fd724a1d515e5cbdeb77ffae96a617a33666656250cfb" exitCode=0 Oct 04 04:48:08 crc kubenswrapper[4992]: I1004 04:48:08.957988 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" event={"ID":"ca142652-0451-4c15-bd36-d2fc14a916ad","Type":"ContainerDied","Data":"8beaa871fe6518c9635fd724a1d515e5cbdeb77ffae96a617a33666656250cfb"} Oct 04 04:48:10 crc kubenswrapper[4992]: I1004 04:48:10.971904 4992 generic.go:334] "Generic (PLEG): container finished" podID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerID="8385462adcaaa109b2af41f2f4aacd007c1fc2c20d95a7f99cde6a003d85aa3d" exitCode=0 Oct 04 04:48:10 crc kubenswrapper[4992]: I1004 04:48:10.971966 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" event={"ID":"ca142652-0451-4c15-bd36-d2fc14a916ad","Type":"ContainerDied","Data":"8385462adcaaa109b2af41f2f4aacd007c1fc2c20d95a7f99cde6a003d85aa3d"} Oct 04 04:48:11 crc kubenswrapper[4992]: I1004 04:48:11.984859 4992 generic.go:334] "Generic (PLEG): container finished" podID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerID="21954075ac768e9aca9bf20c536da30337632e543d24f2eb125d5cd26209a9c9" exitCode=0 Oct 04 04:48:11 crc kubenswrapper[4992]: I1004 04:48:11.985325 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" event={"ID":"ca142652-0451-4c15-bd36-d2fc14a916ad","Type":"ContainerDied","Data":"21954075ac768e9aca9bf20c536da30337632e543d24f2eb125d5cd26209a9c9"} Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.232534 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.399333 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjwt9\" (UniqueName: \"kubernetes.io/projected/ca142652-0451-4c15-bd36-d2fc14a916ad-kube-api-access-kjwt9\") pod \"ca142652-0451-4c15-bd36-d2fc14a916ad\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.399468 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-bundle\") pod \"ca142652-0451-4c15-bd36-d2fc14a916ad\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.399553 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-util\") pod \"ca142652-0451-4c15-bd36-d2fc14a916ad\" (UID: \"ca142652-0451-4c15-bd36-d2fc14a916ad\") " Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.401660 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-bundle" (OuterVolumeSpecName: "bundle") pod "ca142652-0451-4c15-bd36-d2fc14a916ad" (UID: "ca142652-0451-4c15-bd36-d2fc14a916ad"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.407991 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca142652-0451-4c15-bd36-d2fc14a916ad-kube-api-access-kjwt9" (OuterVolumeSpecName: "kube-api-access-kjwt9") pod "ca142652-0451-4c15-bd36-d2fc14a916ad" (UID: "ca142652-0451-4c15-bd36-d2fc14a916ad"). InnerVolumeSpecName "kube-api-access-kjwt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.429561 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-util" (OuterVolumeSpecName: "util") pod "ca142652-0451-4c15-bd36-d2fc14a916ad" (UID: "ca142652-0451-4c15-bd36-d2fc14a916ad"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.500810 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjwt9\" (UniqueName: \"kubernetes.io/projected/ca142652-0451-4c15-bd36-d2fc14a916ad-kube-api-access-kjwt9\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.500864 4992 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:13 crc kubenswrapper[4992]: I1004 04:48:13.500887 4992 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ca142652-0451-4c15-bd36-d2fc14a916ad-util\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:14 crc kubenswrapper[4992]: I1004 04:48:14.000050 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" event={"ID":"ca142652-0451-4c15-bd36-d2fc14a916ad","Type":"ContainerDied","Data":"9b1fb3c24e3df03251c5c40bffe0c4f2431a1051d00d6ac6e356285f1b8e2e1d"} Oct 04 04:48:14 crc kubenswrapper[4992]: I1004 04:48:14.000104 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b1fb3c24e3df03251c5c40bffe0c4f2431a1051d00d6ac6e356285f1b8e2e1d" Oct 04 04:48:14 crc kubenswrapper[4992]: I1004 04:48:14.000193 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.913883 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd"] Oct 04 04:48:26 crc kubenswrapper[4992]: E1004 04:48:26.914582 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerName="pull" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.914595 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerName="pull" Oct 04 04:48:26 crc kubenswrapper[4992]: E1004 04:48:26.914605 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerName="extract" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.914610 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerName="extract" Oct 04 04:48:26 crc kubenswrapper[4992]: E1004 04:48:26.914622 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerName="util" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.914629 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerName="util" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.914717 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca142652-0451-4c15-bd36-d2fc14a916ad" containerName="extract" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.915074 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.921970 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.922031 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.922098 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-xbkf4" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.922472 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.925096 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.935498 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd"] Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.988113 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8fb02cec-1648-47c4-a2af-034fa5f204ca-apiservice-cert\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.988204 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clzvg\" (UniqueName: \"kubernetes.io/projected/8fb02cec-1648-47c4-a2af-034fa5f204ca-kube-api-access-clzvg\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:26 crc kubenswrapper[4992]: I1004 04:48:26.988236 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8fb02cec-1648-47c4-a2af-034fa5f204ca-webhook-cert\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.089868 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8fb02cec-1648-47c4-a2af-034fa5f204ca-apiservice-cert\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.089966 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clzvg\" (UniqueName: \"kubernetes.io/projected/8fb02cec-1648-47c4-a2af-034fa5f204ca-kube-api-access-clzvg\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.090004 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8fb02cec-1648-47c4-a2af-034fa5f204ca-webhook-cert\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.094983 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8fb02cec-1648-47c4-a2af-034fa5f204ca-webhook-cert\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.099866 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8fb02cec-1648-47c4-a2af-034fa5f204ca-apiservice-cert\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.105928 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clzvg\" (UniqueName: \"kubernetes.io/projected/8fb02cec-1648-47c4-a2af-034fa5f204ca-kube-api-access-clzvg\") pod \"metallb-operator-controller-manager-5fb67947c7-9kpxd\" (UID: \"8fb02cec-1648-47c4-a2af-034fa5f204ca\") " pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.157895 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d"] Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.158592 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.161145 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.161609 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-x7gct" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.166704 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.212518 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d"] Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.233000 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.292742 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffthc\" (UniqueName: \"kubernetes.io/projected/f5ef1e00-0f56-41fb-8961-1a0638414444-kube-api-access-ffthc\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.293090 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f5ef1e00-0f56-41fb-8961-1a0638414444-apiservice-cert\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.293266 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f5ef1e00-0f56-41fb-8961-1a0638414444-webhook-cert\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.395283 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f5ef1e00-0f56-41fb-8961-1a0638414444-apiservice-cert\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.395349 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f5ef1e00-0f56-41fb-8961-1a0638414444-webhook-cert\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.395418 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffthc\" (UniqueName: \"kubernetes.io/projected/f5ef1e00-0f56-41fb-8961-1a0638414444-kube-api-access-ffthc\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.405835 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f5ef1e00-0f56-41fb-8961-1a0638414444-apiservice-cert\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.425690 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f5ef1e00-0f56-41fb-8961-1a0638414444-webhook-cert\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.432207 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffthc\" (UniqueName: \"kubernetes.io/projected/f5ef1e00-0f56-41fb-8961-1a0638414444-kube-api-access-ffthc\") pod \"metallb-operator-webhook-server-8b8bf6765-q6l5d\" (UID: \"f5ef1e00-0f56-41fb-8961-1a0638414444\") " pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.452218 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd"] Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.474972 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:27 crc kubenswrapper[4992]: I1004 04:48:27.753576 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d"] Oct 04 04:48:27 crc kubenswrapper[4992]: W1004 04:48:27.761385 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5ef1e00_0f56_41fb_8961_1a0638414444.slice/crio-1ce63a2ab7dde5eb798d50b16962cedd740ae3fa01a960d39f7762f300579268 WatchSource:0}: Error finding container 1ce63a2ab7dde5eb798d50b16962cedd740ae3fa01a960d39f7762f300579268: Status 404 returned error can't find the container with id 1ce63a2ab7dde5eb798d50b16962cedd740ae3fa01a960d39f7762f300579268 Oct 04 04:48:28 crc kubenswrapper[4992]: I1004 04:48:28.078064 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" event={"ID":"f5ef1e00-0f56-41fb-8961-1a0638414444","Type":"ContainerStarted","Data":"1ce63a2ab7dde5eb798d50b16962cedd740ae3fa01a960d39f7762f300579268"} Oct 04 04:48:28 crc kubenswrapper[4992]: I1004 04:48:28.079110 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" event={"ID":"8fb02cec-1648-47c4-a2af-034fa5f204ca","Type":"ContainerStarted","Data":"f552cbb6747f3f633fe7789193838436a86b8cf03ec711e46ee37475cc568011"} Oct 04 04:48:33 crc kubenswrapper[4992]: I1004 04:48:33.110290 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" event={"ID":"f5ef1e00-0f56-41fb-8961-1a0638414444","Type":"ContainerStarted","Data":"ed4465473fa100be8d9d9f7b4eab4df0081e5549e92f6ea73ee8802a94e4ba9a"} Oct 04 04:48:33 crc kubenswrapper[4992]: I1004 04:48:33.110629 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:33 crc kubenswrapper[4992]: I1004 04:48:33.112090 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" event={"ID":"8fb02cec-1648-47c4-a2af-034fa5f204ca","Type":"ContainerStarted","Data":"eae06f88ffb37a3e111a45f7602f4602fbc70b3e8c3c5e8d688915db35faf455"} Oct 04 04:48:33 crc kubenswrapper[4992]: I1004 04:48:33.112605 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:48:33 crc kubenswrapper[4992]: I1004 04:48:33.130779 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" podStartSLOduration=1.8132850980000002 podStartE2EDuration="6.13075533s" podCreationTimestamp="2025-10-04 04:48:27 +0000 UTC" firstStartedPulling="2025-10-04 04:48:27.765414345 +0000 UTC m=+941.613089813" lastFinishedPulling="2025-10-04 04:48:32.082884577 +0000 UTC m=+945.930560045" observedRunningTime="2025-10-04 04:48:33.126812025 +0000 UTC m=+946.974487503" watchObservedRunningTime="2025-10-04 04:48:33.13075533 +0000 UTC m=+946.978430818" Oct 04 04:48:33 crc kubenswrapper[4992]: I1004 04:48:33.149004 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" podStartSLOduration=2.544739821 podStartE2EDuration="7.148974855s" podCreationTimestamp="2025-10-04 04:48:26 +0000 UTC" firstStartedPulling="2025-10-04 04:48:27.462460612 +0000 UTC m=+941.310136080" lastFinishedPulling="2025-10-04 04:48:32.066695646 +0000 UTC m=+945.914371114" observedRunningTime="2025-10-04 04:48:33.145563254 +0000 UTC m=+946.993238762" watchObservedRunningTime="2025-10-04 04:48:33.148974855 +0000 UTC m=+946.996650363" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.070315 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zzngl"] Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.074945 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.086811 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzngl"] Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.205567 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw64m\" (UniqueName: \"kubernetes.io/projected/942b8281-c6b2-45eb-bde6-c7aa1297ace4-kube-api-access-gw64m\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.205618 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-utilities\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.205659 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-catalog-content\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.307142 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw64m\" (UniqueName: \"kubernetes.io/projected/942b8281-c6b2-45eb-bde6-c7aa1297ace4-kube-api-access-gw64m\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.307209 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-utilities\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.307271 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-catalog-content\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.307714 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-utilities\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.307796 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-catalog-content\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.331325 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw64m\" (UniqueName: \"kubernetes.io/projected/942b8281-c6b2-45eb-bde6-c7aa1297ace4-kube-api-access-gw64m\") pod \"certified-operators-zzngl\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.395638 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:42 crc kubenswrapper[4992]: I1004 04:48:42.882610 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzngl"] Oct 04 04:48:43 crc kubenswrapper[4992]: I1004 04:48:43.164867 4992 generic.go:334] "Generic (PLEG): container finished" podID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerID="8002a9c48bde9de6be563160aeddbe0caff69b1119f9bb4ddbec4a205a4ddbfe" exitCode=0 Oct 04 04:48:43 crc kubenswrapper[4992]: I1004 04:48:43.164930 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzngl" event={"ID":"942b8281-c6b2-45eb-bde6-c7aa1297ace4","Type":"ContainerDied","Data":"8002a9c48bde9de6be563160aeddbe0caff69b1119f9bb4ddbec4a205a4ddbfe"} Oct 04 04:48:43 crc kubenswrapper[4992]: I1004 04:48:43.165201 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzngl" event={"ID":"942b8281-c6b2-45eb-bde6-c7aa1297ace4","Type":"ContainerStarted","Data":"953144e9ec30415bf60b98adfd3774a8cb32c36de78680fc33c10f4cd7743e06"} Oct 04 04:48:44 crc kubenswrapper[4992]: I1004 04:48:44.173783 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzngl" event={"ID":"942b8281-c6b2-45eb-bde6-c7aa1297ace4","Type":"ContainerStarted","Data":"451756a922bfef8683a4728dded6ed76ffd4b5a210c0ddd7d180d0c2981d9647"} Oct 04 04:48:45 crc kubenswrapper[4992]: I1004 04:48:45.181671 4992 generic.go:334] "Generic (PLEG): container finished" podID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerID="451756a922bfef8683a4728dded6ed76ffd4b5a210c0ddd7d180d0c2981d9647" exitCode=0 Oct 04 04:48:45 crc kubenswrapper[4992]: I1004 04:48:45.181753 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzngl" event={"ID":"942b8281-c6b2-45eb-bde6-c7aa1297ace4","Type":"ContainerDied","Data":"451756a922bfef8683a4728dded6ed76ffd4b5a210c0ddd7d180d0c2981d9647"} Oct 04 04:48:46 crc kubenswrapper[4992]: I1004 04:48:46.189262 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzngl" event={"ID":"942b8281-c6b2-45eb-bde6-c7aa1297ace4","Type":"ContainerStarted","Data":"e1cc9bff304bbcc7ab996479638a36c79c9700261c37ec1c898659c999efd530"} Oct 04 04:48:46 crc kubenswrapper[4992]: I1004 04:48:46.218925 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zzngl" podStartSLOduration=1.68638214 podStartE2EDuration="4.21890795s" podCreationTimestamp="2025-10-04 04:48:42 +0000 UTC" firstStartedPulling="2025-10-04 04:48:43.166117661 +0000 UTC m=+957.013793129" lastFinishedPulling="2025-10-04 04:48:45.698643471 +0000 UTC m=+959.546318939" observedRunningTime="2025-10-04 04:48:46.214937984 +0000 UTC m=+960.062613452" watchObservedRunningTime="2025-10-04 04:48:46.21890795 +0000 UTC m=+960.066583418" Oct 04 04:48:47 crc kubenswrapper[4992]: I1004 04:48:47.479292 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-8b8bf6765-q6l5d" Oct 04 04:48:52 crc kubenswrapper[4992]: I1004 04:48:52.396448 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:52 crc kubenswrapper[4992]: I1004 04:48:52.396775 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:52 crc kubenswrapper[4992]: I1004 04:48:52.436798 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:53 crc kubenswrapper[4992]: I1004 04:48:53.265493 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:54 crc kubenswrapper[4992]: I1004 04:48:54.649152 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzngl"] Oct 04 04:48:55 crc kubenswrapper[4992]: I1004 04:48:55.237015 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zzngl" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerName="registry-server" containerID="cri-o://e1cc9bff304bbcc7ab996479638a36c79c9700261c37ec1c898659c999efd530" gracePeriod=2 Oct 04 04:48:56 crc kubenswrapper[4992]: I1004 04:48:56.248837 4992 generic.go:334] "Generic (PLEG): container finished" podID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerID="e1cc9bff304bbcc7ab996479638a36c79c9700261c37ec1c898659c999efd530" exitCode=0 Oct 04 04:48:56 crc kubenswrapper[4992]: I1004 04:48:56.248937 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzngl" event={"ID":"942b8281-c6b2-45eb-bde6-c7aa1297ace4","Type":"ContainerDied","Data":"e1cc9bff304bbcc7ab996479638a36c79c9700261c37ec1c898659c999efd530"} Oct 04 04:48:56 crc kubenswrapper[4992]: I1004 04:48:56.849087 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.013321 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-utilities\") pod \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.013961 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-catalog-content\") pod \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.014000 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw64m\" (UniqueName: \"kubernetes.io/projected/942b8281-c6b2-45eb-bde6-c7aa1297ace4-kube-api-access-gw64m\") pod \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\" (UID: \"942b8281-c6b2-45eb-bde6-c7aa1297ace4\") " Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.014543 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-utilities" (OuterVolumeSpecName: "utilities") pod "942b8281-c6b2-45eb-bde6-c7aa1297ace4" (UID: "942b8281-c6b2-45eb-bde6-c7aa1297ace4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.022160 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/942b8281-c6b2-45eb-bde6-c7aa1297ace4-kube-api-access-gw64m" (OuterVolumeSpecName: "kube-api-access-gw64m") pod "942b8281-c6b2-45eb-bde6-c7aa1297ace4" (UID: "942b8281-c6b2-45eb-bde6-c7aa1297ace4"). InnerVolumeSpecName "kube-api-access-gw64m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.086007 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "942b8281-c6b2-45eb-bde6-c7aa1297ace4" (UID: "942b8281-c6b2-45eb-bde6-c7aa1297ace4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.115792 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.116144 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/942b8281-c6b2-45eb-bde6-c7aa1297ace4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.116288 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw64m\" (UniqueName: \"kubernetes.io/projected/942b8281-c6b2-45eb-bde6-c7aa1297ace4-kube-api-access-gw64m\") on node \"crc\" DevicePath \"\"" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.258915 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzngl" event={"ID":"942b8281-c6b2-45eb-bde6-c7aa1297ace4","Type":"ContainerDied","Data":"953144e9ec30415bf60b98adfd3774a8cb32c36de78680fc33c10f4cd7743e06"} Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.258986 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzngl" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.258996 4992 scope.go:117] "RemoveContainer" containerID="e1cc9bff304bbcc7ab996479638a36c79c9700261c37ec1c898659c999efd530" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.280163 4992 scope.go:117] "RemoveContainer" containerID="451756a922bfef8683a4728dded6ed76ffd4b5a210c0ddd7d180d0c2981d9647" Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.295701 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzngl"] Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.302437 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zzngl"] Oct 04 04:48:57 crc kubenswrapper[4992]: I1004 04:48:57.315677 4992 scope.go:117] "RemoveContainer" containerID="8002a9c48bde9de6be563160aeddbe0caff69b1119f9bb4ddbec4a205a4ddbfe" Oct 04 04:48:58 crc kubenswrapper[4992]: I1004 04:48:58.327449 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" path="/var/lib/kubelet/pods/942b8281-c6b2-45eb-bde6-c7aa1297ace4/volumes" Oct 04 04:49:07 crc kubenswrapper[4992]: I1004 04:49:07.237529 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5fb67947c7-9kpxd" Oct 04 04:49:07 crc kubenswrapper[4992]: I1004 04:49:07.997840 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-lb5zk"] Oct 04 04:49:07 crc kubenswrapper[4992]: E1004 04:49:07.998480 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerName="registry-server" Oct 04 04:49:07 crc kubenswrapper[4992]: I1004 04:49:07.998496 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerName="registry-server" Oct 04 04:49:07 crc kubenswrapper[4992]: E1004 04:49:07.998515 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerName="extract-content" Oct 04 04:49:07 crc kubenswrapper[4992]: I1004 04:49:07.998523 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerName="extract-content" Oct 04 04:49:07 crc kubenswrapper[4992]: E1004 04:49:07.998540 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerName="extract-utilities" Oct 04 04:49:07 crc kubenswrapper[4992]: I1004 04:49:07.998548 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerName="extract-utilities" Oct 04 04:49:07 crc kubenswrapper[4992]: I1004 04:49:07.998698 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="942b8281-c6b2-45eb-bde6-c7aa1297ace4" containerName="registry-server" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.001091 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.002026 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7"] Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.002898 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.006757 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.006757 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.006754 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.009046 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-szk67" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.019004 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7"] Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.108845 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-nbkqc"] Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.109682 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.111321 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.112683 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-wvk2f" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.112745 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.112837 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.118672 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-lkqgp"] Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.120064 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.122039 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.131302 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-lkqgp"] Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163266 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-conf\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163313 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-866zz\" (UniqueName: \"kubernetes.io/projected/64b51eb5-0b5d-4cc2-ada8-889a67608f62-kube-api-access-866zz\") pod \"frr-k8s-webhook-server-64bf5d555-8tns7\" (UID: \"64b51eb5-0b5d-4cc2-ada8-889a67608f62\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163338 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64b51eb5-0b5d-4cc2-ada8-889a67608f62-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8tns7\" (UID: \"64b51eb5-0b5d-4cc2-ada8-889a67608f62\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163477 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-metrics\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163573 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-sockets\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163597 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-metrics-certs\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163624 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-startup\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163712 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46xbj\" (UniqueName: \"kubernetes.io/projected/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-kube-api-access-46xbj\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.163776 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-reloader\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264764 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-metrics-certs\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264806 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-sockets\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264823 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-startup\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264857 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3f777561-a01c-4e77-98f4-eb7b15f56eca-cert\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264883 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-metrics-certs\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264902 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46xbj\" (UniqueName: \"kubernetes.io/projected/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-kube-api-access-46xbj\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264919 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfl24\" (UniqueName: \"kubernetes.io/projected/3f777561-a01c-4e77-98f4-eb7b15f56eca-kube-api-access-hfl24\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264949 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-reloader\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264966 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/17e67a6f-46b1-434e-a422-21f96bf870ea-metallb-excludel2\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.264983 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-conf\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.265000 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f777561-a01c-4e77-98f4-eb7b15f56eca-metrics-certs\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.265018 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-866zz\" (UniqueName: \"kubernetes.io/projected/64b51eb5-0b5d-4cc2-ada8-889a67608f62-kube-api-access-866zz\") pod \"frr-k8s-webhook-server-64bf5d555-8tns7\" (UID: \"64b51eb5-0b5d-4cc2-ada8-889a67608f62\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.265035 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64b51eb5-0b5d-4cc2-ada8-889a67608f62-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8tns7\" (UID: \"64b51eb5-0b5d-4cc2-ada8-889a67608f62\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.265052 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.265074 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-metrics\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.265100 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxm8g\" (UniqueName: \"kubernetes.io/projected/17e67a6f-46b1-434e-a422-21f96bf870ea-kube-api-access-pxm8g\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.265574 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-sockets\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: E1004 04:49:08.266028 4992 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 04 04:49:08 crc kubenswrapper[4992]: E1004 04:49:08.266117 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64b51eb5-0b5d-4cc2-ada8-889a67608f62-cert podName:64b51eb5-0b5d-4cc2-ada8-889a67608f62 nodeName:}" failed. No retries permitted until 2025-10-04 04:49:08.766093276 +0000 UTC m=+982.613768744 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64b51eb5-0b5d-4cc2-ada8-889a67608f62-cert") pod "frr-k8s-webhook-server-64bf5d555-8tns7" (UID: "64b51eb5-0b5d-4cc2-ada8-889a67608f62") : secret "frr-k8s-webhook-server-cert" not found Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.266486 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-reloader\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.266520 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-startup\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.266636 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-metrics\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.266741 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-frr-conf\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.274376 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-metrics-certs\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.286444 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46xbj\" (UniqueName: \"kubernetes.io/projected/6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1-kube-api-access-46xbj\") pod \"frr-k8s-lb5zk\" (UID: \"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1\") " pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.292873 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-866zz\" (UniqueName: \"kubernetes.io/projected/64b51eb5-0b5d-4cc2-ada8-889a67608f62-kube-api-access-866zz\") pod \"frr-k8s-webhook-server-64bf5d555-8tns7\" (UID: \"64b51eb5-0b5d-4cc2-ada8-889a67608f62\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.319931 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.365996 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxm8g\" (UniqueName: \"kubernetes.io/projected/17e67a6f-46b1-434e-a422-21f96bf870ea-kube-api-access-pxm8g\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.366065 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3f777561-a01c-4e77-98f4-eb7b15f56eca-cert\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.366091 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-metrics-certs\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.366111 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfl24\" (UniqueName: \"kubernetes.io/projected/3f777561-a01c-4e77-98f4-eb7b15f56eca-kube-api-access-hfl24\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.366142 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/17e67a6f-46b1-434e-a422-21f96bf870ea-metallb-excludel2\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.366157 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f777561-a01c-4e77-98f4-eb7b15f56eca-metrics-certs\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.366184 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: E1004 04:49:08.366281 4992 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 04:49:08 crc kubenswrapper[4992]: E1004 04:49:08.366328 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist podName:17e67a6f-46b1-434e-a422-21f96bf870ea nodeName:}" failed. No retries permitted until 2025-10-04 04:49:08.86631403 +0000 UTC m=+982.713989498 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist") pod "speaker-nbkqc" (UID: "17e67a6f-46b1-434e-a422-21f96bf870ea") : secret "metallb-memberlist" not found Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.367568 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/17e67a6f-46b1-434e-a422-21f96bf870ea-metallb-excludel2\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.370589 4992 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.371136 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f777561-a01c-4e77-98f4-eb7b15f56eca-metrics-certs\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.373006 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-metrics-certs\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.380864 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3f777561-a01c-4e77-98f4-eb7b15f56eca-cert\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.387612 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfl24\" (UniqueName: \"kubernetes.io/projected/3f777561-a01c-4e77-98f4-eb7b15f56eca-kube-api-access-hfl24\") pod \"controller-68d546b9d8-lkqgp\" (UID: \"3f777561-a01c-4e77-98f4-eb7b15f56eca\") " pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.388546 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxm8g\" (UniqueName: \"kubernetes.io/projected/17e67a6f-46b1-434e-a422-21f96bf870ea-kube-api-access-pxm8g\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.433288 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.777025 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64b51eb5-0b5d-4cc2-ada8-889a67608f62-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8tns7\" (UID: \"64b51eb5-0b5d-4cc2-ada8-889a67608f62\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.781528 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64b51eb5-0b5d-4cc2-ada8-889a67608f62-cert\") pod \"frr-k8s-webhook-server-64bf5d555-8tns7\" (UID: \"64b51eb5-0b5d-4cc2-ada8-889a67608f62\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.831737 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-lkqgp"] Oct 04 04:49:08 crc kubenswrapper[4992]: W1004 04:49:08.836911 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f777561_a01c_4e77_98f4_eb7b15f56eca.slice/crio-f0233355ab2c6bc5a0d66c6f84c9c53966c9fbd9529b84cf01a1ba8f0c7c21fd WatchSource:0}: Error finding container f0233355ab2c6bc5a0d66c6f84c9c53966c9fbd9529b84cf01a1ba8f0c7c21fd: Status 404 returned error can't find the container with id f0233355ab2c6bc5a0d66c6f84c9c53966c9fbd9529b84cf01a1ba8f0c7c21fd Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.878254 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:08 crc kubenswrapper[4992]: E1004 04:49:08.878569 4992 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 04:49:08 crc kubenswrapper[4992]: E1004 04:49:08.878667 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist podName:17e67a6f-46b1-434e-a422-21f96bf870ea nodeName:}" failed. No retries permitted until 2025-10-04 04:49:09.878639482 +0000 UTC m=+983.726314990 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist") pod "speaker-nbkqc" (UID: "17e67a6f-46b1-434e-a422-21f96bf870ea") : secret "metallb-memberlist" not found Oct 04 04:49:08 crc kubenswrapper[4992]: I1004 04:49:08.929720 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.138122 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7"] Oct 04 04:49:09 crc kubenswrapper[4992]: W1004 04:49:09.140836 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64b51eb5_0b5d_4cc2_ada8_889a67608f62.slice/crio-361f1ff91c604da5859ce8a24552b767089f7132c36503f5fc60f9f5040fd6db WatchSource:0}: Error finding container 361f1ff91c604da5859ce8a24552b767089f7132c36503f5fc60f9f5040fd6db: Status 404 returned error can't find the container with id 361f1ff91c604da5859ce8a24552b767089f7132c36503f5fc60f9f5040fd6db Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.321599 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-lkqgp" event={"ID":"3f777561-a01c-4e77-98f4-eb7b15f56eca","Type":"ContainerStarted","Data":"1b29db1d34aa6d7f4f7a17c24332ca64f570f5e68f1f9160fe9d1db55cdede93"} Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.321823 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-lkqgp" event={"ID":"3f777561-a01c-4e77-98f4-eb7b15f56eca","Type":"ContainerStarted","Data":"110b7209d4cb1ecaee1795f8511a118614878fb49493fc70b462387689506ad1"} Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.321843 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.321854 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-lkqgp" event={"ID":"3f777561-a01c-4e77-98f4-eb7b15f56eca","Type":"ContainerStarted","Data":"f0233355ab2c6bc5a0d66c6f84c9c53966c9fbd9529b84cf01a1ba8f0c7c21fd"} Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.324062 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" event={"ID":"64b51eb5-0b5d-4cc2-ada8-889a67608f62","Type":"ContainerStarted","Data":"361f1ff91c604da5859ce8a24552b767089f7132c36503f5fc60f9f5040fd6db"} Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.325257 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerStarted","Data":"35fcf072978243cf76b27f738832c3fb19fde071a77bbf2c69e045f4bd01f8fa"} Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.337424 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-lkqgp" podStartSLOduration=1.337405157 podStartE2EDuration="1.337405157s" podCreationTimestamp="2025-10-04 04:49:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:49:09.333142446 +0000 UTC m=+983.180817914" watchObservedRunningTime="2025-10-04 04:49:09.337405157 +0000 UTC m=+983.185080625" Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.921987 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:09 crc kubenswrapper[4992]: I1004 04:49:09.928103 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/17e67a6f-46b1-434e-a422-21f96bf870ea-memberlist\") pod \"speaker-nbkqc\" (UID: \"17e67a6f-46b1-434e-a422-21f96bf870ea\") " pod="metallb-system/speaker-nbkqc" Oct 04 04:49:10 crc kubenswrapper[4992]: I1004 04:49:10.222617 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-nbkqc" Oct 04 04:49:10 crc kubenswrapper[4992]: I1004 04:49:10.366916 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nbkqc" event={"ID":"17e67a6f-46b1-434e-a422-21f96bf870ea","Type":"ContainerStarted","Data":"87d7a901e27cbf6150aa90875212656f46a7839b29fd6ce45a9c3cc07fd7e876"} Oct 04 04:49:11 crc kubenswrapper[4992]: I1004 04:49:11.376265 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nbkqc" event={"ID":"17e67a6f-46b1-434e-a422-21f96bf870ea","Type":"ContainerStarted","Data":"5884635780d371bee166b931bef139bf2db2eaaa9fa94efba4d57fdfcfdc3160"} Oct 04 04:49:11 crc kubenswrapper[4992]: I1004 04:49:11.376647 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-nbkqc" Oct 04 04:49:11 crc kubenswrapper[4992]: I1004 04:49:11.376662 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-nbkqc" event={"ID":"17e67a6f-46b1-434e-a422-21f96bf870ea","Type":"ContainerStarted","Data":"76cf9da0d783037040edfb32c92f0e5466fd846b6229c331a7284161103eca17"} Oct 04 04:49:11 crc kubenswrapper[4992]: I1004 04:49:11.395843 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-nbkqc" podStartSLOduration=3.395827793 podStartE2EDuration="3.395827793s" podCreationTimestamp="2025-10-04 04:49:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:49:11.392429904 +0000 UTC m=+985.240105372" watchObservedRunningTime="2025-10-04 04:49:11.395827793 +0000 UTC m=+985.243503261" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.423923 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zg968"] Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.425099 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.427657 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg968"] Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.471977 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-catalog-content\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.472055 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4c77\" (UniqueName: \"kubernetes.io/projected/82bcfc12-3f52-4014-a93e-98c05d64860f-kube-api-access-j4c77\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.472102 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-utilities\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.573293 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-utilities\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.573375 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-catalog-content\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.573419 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4c77\" (UniqueName: \"kubernetes.io/projected/82bcfc12-3f52-4014-a93e-98c05d64860f-kube-api-access-j4c77\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.574053 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-utilities\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.574248 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-catalog-content\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.589987 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4c77\" (UniqueName: \"kubernetes.io/projected/82bcfc12-3f52-4014-a93e-98c05d64860f-kube-api-access-j4c77\") pod \"redhat-marketplace-zg968\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:13 crc kubenswrapper[4992]: I1004 04:49:13.746183 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:17 crc kubenswrapper[4992]: I1004 04:49:17.510288 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg968"] Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.418256 4992 generic.go:334] "Generic (PLEG): container finished" podID="6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1" containerID="ba5966cb7b55bf1a941585678981b4a49f174efd762bde326413690324334aa1" exitCode=0 Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.418413 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerDied","Data":"ba5966cb7b55bf1a941585678981b4a49f174efd762bde326413690324334aa1"} Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.419993 4992 generic.go:334] "Generic (PLEG): container finished" podID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerID="c41cce470a59e81cc8d21a264f22178981ee227f5e881c524d4aee4ff9f3e1b0" exitCode=0 Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.420028 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg968" event={"ID":"82bcfc12-3f52-4014-a93e-98c05d64860f","Type":"ContainerDied","Data":"c41cce470a59e81cc8d21a264f22178981ee227f5e881c524d4aee4ff9f3e1b0"} Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.420045 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg968" event={"ID":"82bcfc12-3f52-4014-a93e-98c05d64860f","Type":"ContainerStarted","Data":"298a2bbda5a68d06202b019cbdac6bfd86d88506c4d0bb4251a2ad2ddffa9407"} Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.422037 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" event={"ID":"64b51eb5-0b5d-4cc2-ada8-889a67608f62","Type":"ContainerStarted","Data":"0fe94658fe48ee06320c5e34e599a138086ab783165ec7c67708bf7f91a557fb"} Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.422221 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.438214 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-lkqgp" Oct 04 04:49:18 crc kubenswrapper[4992]: I1004 04:49:18.579124 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" podStartSLOduration=2.856944998 podStartE2EDuration="11.579092552s" podCreationTimestamp="2025-10-04 04:49:07 +0000 UTC" firstStartedPulling="2025-10-04 04:49:09.144004223 +0000 UTC m=+982.991679691" lastFinishedPulling="2025-10-04 04:49:17.866151777 +0000 UTC m=+991.713827245" observedRunningTime="2025-10-04 04:49:18.538437231 +0000 UTC m=+992.386112699" watchObservedRunningTime="2025-10-04 04:49:18.579092552 +0000 UTC m=+992.426768020" Oct 04 04:49:19 crc kubenswrapper[4992]: I1004 04:49:19.428809 4992 generic.go:334] "Generic (PLEG): container finished" podID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerID="f41ba888c3a839b392b81ddfa0f1e46f9103df1ea4785abba6c1d4fcfde68c2f" exitCode=0 Oct 04 04:49:19 crc kubenswrapper[4992]: I1004 04:49:19.428902 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg968" event={"ID":"82bcfc12-3f52-4014-a93e-98c05d64860f","Type":"ContainerDied","Data":"f41ba888c3a839b392b81ddfa0f1e46f9103df1ea4785abba6c1d4fcfde68c2f"} Oct 04 04:49:19 crc kubenswrapper[4992]: I1004 04:49:19.432627 4992 generic.go:334] "Generic (PLEG): container finished" podID="6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1" containerID="43e009757e270e790b06abbe536a85fb9d66e233436e04d135df0a80bf608bfd" exitCode=0 Oct 04 04:49:19 crc kubenswrapper[4992]: I1004 04:49:19.432672 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerDied","Data":"43e009757e270e790b06abbe536a85fb9d66e233436e04d135df0a80bf608bfd"} Oct 04 04:49:20 crc kubenswrapper[4992]: I1004 04:49:20.227039 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-nbkqc" Oct 04 04:49:20 crc kubenswrapper[4992]: I1004 04:49:20.439904 4992 generic.go:334] "Generic (PLEG): container finished" podID="6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1" containerID="95557cefec82dc525d2b9bde830939bd343ae3450aa340756b566ad79328f1e2" exitCode=0 Oct 04 04:49:20 crc kubenswrapper[4992]: I1004 04:49:20.440001 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerDied","Data":"95557cefec82dc525d2b9bde830939bd343ae3450aa340756b566ad79328f1e2"} Oct 04 04:49:20 crc kubenswrapper[4992]: I1004 04:49:20.442436 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg968" event={"ID":"82bcfc12-3f52-4014-a93e-98c05d64860f","Type":"ContainerStarted","Data":"437c8cb64151c5ac589541e201f39529ff462ff2cd55359cdc4eda0aa3880fe7"} Oct 04 04:49:20 crc kubenswrapper[4992]: I1004 04:49:20.481639 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zg968" podStartSLOduration=5.924706046 podStartE2EDuration="7.481619043s" podCreationTimestamp="2025-10-04 04:49:13 +0000 UTC" firstStartedPulling="2025-10-04 04:49:18.421169073 +0000 UTC m=+992.268844541" lastFinishedPulling="2025-10-04 04:49:19.97808206 +0000 UTC m=+993.825757538" observedRunningTime="2025-10-04 04:49:20.480115733 +0000 UTC m=+994.327791201" watchObservedRunningTime="2025-10-04 04:49:20.481619043 +0000 UTC m=+994.329294511" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.451951 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerStarted","Data":"b3dc6382b514c8df32c276586ec58bf69410db57da440cf4cfd0edd20f6d2ec4"} Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.587110 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gt8f6"] Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.588228 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.599052 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gt8f6"] Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.785419 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-utilities\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.785497 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pzzf\" (UniqueName: \"kubernetes.io/projected/960e23d5-abd5-46aa-ba53-36c5e91b3a61-kube-api-access-2pzzf\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.785694 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-catalog-content\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.886867 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pzzf\" (UniqueName: \"kubernetes.io/projected/960e23d5-abd5-46aa-ba53-36c5e91b3a61-kube-api-access-2pzzf\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.886985 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-catalog-content\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.887036 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-utilities\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.887622 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-catalog-content\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.887768 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-utilities\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:21 crc kubenswrapper[4992]: I1004 04:49:21.953969 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pzzf\" (UniqueName: \"kubernetes.io/projected/960e23d5-abd5-46aa-ba53-36c5e91b3a61-kube-api-access-2pzzf\") pod \"community-operators-gt8f6\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:22 crc kubenswrapper[4992]: I1004 04:49:22.233268 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:22 crc kubenswrapper[4992]: I1004 04:49:22.462853 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerStarted","Data":"a088643a79b365ac326b0acc3029390244e619568adb0b4a2ec78392d7627478"} Oct 04 04:49:22 crc kubenswrapper[4992]: I1004 04:49:22.463206 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerStarted","Data":"19eb2affa082877d0484a8f63bef1ddb15c0c52db00168281405632b6ed9bff7"} Oct 04 04:49:22 crc kubenswrapper[4992]: I1004 04:49:22.463221 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerStarted","Data":"45d1d178d0043754ed933415728216879558dc71a8f46efff8ebfcaeab5f5595"} Oct 04 04:49:22 crc kubenswrapper[4992]: I1004 04:49:22.629051 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gt8f6"] Oct 04 04:49:22 crc kubenswrapper[4992]: W1004 04:49:22.639601 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod960e23d5_abd5_46aa_ba53_36c5e91b3a61.slice/crio-b8c5b6c9719c2fb116b6728251b7a8e22f83eb2f471685e43872fa05456f86f5 WatchSource:0}: Error finding container b8c5b6c9719c2fb116b6728251b7a8e22f83eb2f471685e43872fa05456f86f5: Status 404 returned error can't find the container with id b8c5b6c9719c2fb116b6728251b7a8e22f83eb2f471685e43872fa05456f86f5 Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.472389 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerStarted","Data":"7a97b6e5b791fcced5d399d2d5c6ecb75d62ab10d689b32aceab5e38185983a6"} Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.472428 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lb5zk" event={"ID":"6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1","Type":"ContainerStarted","Data":"89e9d29655ae25397fa059bb25f84340e8a572ab55613bb6f427cde1f04c94fe"} Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.472541 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.473635 4992 generic.go:334] "Generic (PLEG): container finished" podID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerID="8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898" exitCode=0 Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.473685 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt8f6" event={"ID":"960e23d5-abd5-46aa-ba53-36c5e91b3a61","Type":"ContainerDied","Data":"8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898"} Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.473745 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt8f6" event={"ID":"960e23d5-abd5-46aa-ba53-36c5e91b3a61","Type":"ContainerStarted","Data":"b8c5b6c9719c2fb116b6728251b7a8e22f83eb2f471685e43872fa05456f86f5"} Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.499513 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-lb5zk" podStartSLOduration=7.106843171 podStartE2EDuration="16.499490623s" podCreationTimestamp="2025-10-04 04:49:07 +0000 UTC" firstStartedPulling="2025-10-04 04:49:08.446510331 +0000 UTC m=+982.294185799" lastFinishedPulling="2025-10-04 04:49:17.839157783 +0000 UTC m=+991.686833251" observedRunningTime="2025-10-04 04:49:23.497686665 +0000 UTC m=+997.345362153" watchObservedRunningTime="2025-10-04 04:49:23.499490623 +0000 UTC m=+997.347166091" Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.746637 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.746694 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:23 crc kubenswrapper[4992]: I1004 04:49:23.809264 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:24 crc kubenswrapper[4992]: I1004 04:49:24.482025 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt8f6" event={"ID":"960e23d5-abd5-46aa-ba53-36c5e91b3a61","Type":"ContainerStarted","Data":"7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb"} Oct 04 04:49:25 crc kubenswrapper[4992]: I1004 04:49:25.491308 4992 generic.go:334] "Generic (PLEG): container finished" podID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerID="7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb" exitCode=0 Oct 04 04:49:25 crc kubenswrapper[4992]: I1004 04:49:25.491376 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt8f6" event={"ID":"960e23d5-abd5-46aa-ba53-36c5e91b3a61","Type":"ContainerDied","Data":"7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb"} Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.499673 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt8f6" event={"ID":"960e23d5-abd5-46aa-ba53-36c5e91b3a61","Type":"ContainerStarted","Data":"afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556"} Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.759632 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gt8f6" podStartSLOduration=3.193037202 podStartE2EDuration="5.759617341s" podCreationTimestamp="2025-10-04 04:49:21 +0000 UTC" firstStartedPulling="2025-10-04 04:49:23.474971843 +0000 UTC m=+997.322647311" lastFinishedPulling="2025-10-04 04:49:26.041551982 +0000 UTC m=+999.889227450" observedRunningTime="2025-10-04 04:49:26.516612843 +0000 UTC m=+1000.364288311" watchObservedRunningTime="2025-10-04 04:49:26.759617341 +0000 UTC m=+1000.607292809" Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.763478 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xp6bx"] Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.764318 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.766801 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.766803 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-qzm6q" Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.770746 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.779252 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xp6bx"] Oct 04 04:49:26 crc kubenswrapper[4992]: I1004 04:49:26.951408 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbq88\" (UniqueName: \"kubernetes.io/projected/fd7bfac1-0f3c-4e68-9043-91eabbd07afb-kube-api-access-cbq88\") pod \"openstack-operator-index-xp6bx\" (UID: \"fd7bfac1-0f3c-4e68-9043-91eabbd07afb\") " pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:27 crc kubenswrapper[4992]: I1004 04:49:27.052562 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbq88\" (UniqueName: \"kubernetes.io/projected/fd7bfac1-0f3c-4e68-9043-91eabbd07afb-kube-api-access-cbq88\") pod \"openstack-operator-index-xp6bx\" (UID: \"fd7bfac1-0f3c-4e68-9043-91eabbd07afb\") " pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:27 crc kubenswrapper[4992]: I1004 04:49:27.071939 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbq88\" (UniqueName: \"kubernetes.io/projected/fd7bfac1-0f3c-4e68-9043-91eabbd07afb-kube-api-access-cbq88\") pod \"openstack-operator-index-xp6bx\" (UID: \"fd7bfac1-0f3c-4e68-9043-91eabbd07afb\") " pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:27 crc kubenswrapper[4992]: I1004 04:49:27.082261 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:27 crc kubenswrapper[4992]: I1004 04:49:27.572486 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xp6bx"] Oct 04 04:49:28 crc kubenswrapper[4992]: I1004 04:49:28.334942 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:28 crc kubenswrapper[4992]: I1004 04:49:28.361966 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:28 crc kubenswrapper[4992]: I1004 04:49:28.517467 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xp6bx" event={"ID":"fd7bfac1-0f3c-4e68-9043-91eabbd07afb","Type":"ContainerStarted","Data":"6ca9f132e81f33f4ae8f52ddf97484eefa61caee660c48888f605c56530c3b14"} Oct 04 04:49:28 crc kubenswrapper[4992]: I1004 04:49:28.933751 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-8tns7" Oct 04 04:49:32 crc kubenswrapper[4992]: I1004 04:49:32.234142 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:32 crc kubenswrapper[4992]: I1004 04:49:32.235053 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:32 crc kubenswrapper[4992]: I1004 04:49:32.280024 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:32 crc kubenswrapper[4992]: I1004 04:49:32.540781 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xp6bx" event={"ID":"fd7bfac1-0f3c-4e68-9043-91eabbd07afb","Type":"ContainerStarted","Data":"81d55a4e1c0fd37325b0dcf10a73d5de7da2ca425c6cd440e9eb99bb8c7bf554"} Oct 04 04:49:32 crc kubenswrapper[4992]: I1004 04:49:32.554992 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xp6bx" podStartSLOduration=2.316080715 podStartE2EDuration="6.55497538s" podCreationTimestamp="2025-10-04 04:49:26 +0000 UTC" firstStartedPulling="2025-10-04 04:49:27.585406589 +0000 UTC m=+1001.433082067" lastFinishedPulling="2025-10-04 04:49:31.824301264 +0000 UTC m=+1005.671976732" observedRunningTime="2025-10-04 04:49:32.554917369 +0000 UTC m=+1006.402592867" watchObservedRunningTime="2025-10-04 04:49:32.55497538 +0000 UTC m=+1006.402650868" Oct 04 04:49:32 crc kubenswrapper[4992]: I1004 04:49:32.581390 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:33 crc kubenswrapper[4992]: I1004 04:49:33.791556 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:34 crc kubenswrapper[4992]: I1004 04:49:34.959698 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gt8f6"] Oct 04 04:49:34 crc kubenswrapper[4992]: I1004 04:49:34.959952 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gt8f6" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerName="registry-server" containerID="cri-o://afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556" gracePeriod=2 Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.448980 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.567219 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-utilities\") pod \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.567262 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pzzf\" (UniqueName: \"kubernetes.io/projected/960e23d5-abd5-46aa-ba53-36c5e91b3a61-kube-api-access-2pzzf\") pod \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.567323 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-catalog-content\") pod \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\" (UID: \"960e23d5-abd5-46aa-ba53-36c5e91b3a61\") " Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.567478 4992 generic.go:334] "Generic (PLEG): container finished" podID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerID="afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556" exitCode=0 Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.567515 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt8f6" event={"ID":"960e23d5-abd5-46aa-ba53-36c5e91b3a61","Type":"ContainerDied","Data":"afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556"} Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.567543 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt8f6" event={"ID":"960e23d5-abd5-46aa-ba53-36c5e91b3a61","Type":"ContainerDied","Data":"b8c5b6c9719c2fb116b6728251b7a8e22f83eb2f471685e43872fa05456f86f5"} Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.567562 4992 scope.go:117] "RemoveContainer" containerID="afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.567686 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gt8f6" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.568216 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-utilities" (OuterVolumeSpecName: "utilities") pod "960e23d5-abd5-46aa-ba53-36c5e91b3a61" (UID: "960e23d5-abd5-46aa-ba53-36c5e91b3a61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.573333 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/960e23d5-abd5-46aa-ba53-36c5e91b3a61-kube-api-access-2pzzf" (OuterVolumeSpecName: "kube-api-access-2pzzf") pod "960e23d5-abd5-46aa-ba53-36c5e91b3a61" (UID: "960e23d5-abd5-46aa-ba53-36c5e91b3a61"). InnerVolumeSpecName "kube-api-access-2pzzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.606301 4992 scope.go:117] "RemoveContainer" containerID="7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.621427 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "960e23d5-abd5-46aa-ba53-36c5e91b3a61" (UID: "960e23d5-abd5-46aa-ba53-36c5e91b3a61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.622172 4992 scope.go:117] "RemoveContainer" containerID="8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.643430 4992 scope.go:117] "RemoveContainer" containerID="afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556" Oct 04 04:49:35 crc kubenswrapper[4992]: E1004 04:49:35.643884 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556\": container with ID starting with afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556 not found: ID does not exist" containerID="afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.643940 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556"} err="failed to get container status \"afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556\": rpc error: code = NotFound desc = could not find container \"afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556\": container with ID starting with afafda6ba96c8a9ff0633c768c5b7f7c8cc217d8c8940e0678307d2784974556 not found: ID does not exist" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.643968 4992 scope.go:117] "RemoveContainer" containerID="7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb" Oct 04 04:49:35 crc kubenswrapper[4992]: E1004 04:49:35.644407 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb\": container with ID starting with 7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb not found: ID does not exist" containerID="7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.644440 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb"} err="failed to get container status \"7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb\": rpc error: code = NotFound desc = could not find container \"7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb\": container with ID starting with 7d6b3b9866db52d5e8fb32c7cbb7c157d65f49548622e3605e39103a6bfa06fb not found: ID does not exist" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.644459 4992 scope.go:117] "RemoveContainer" containerID="8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898" Oct 04 04:49:35 crc kubenswrapper[4992]: E1004 04:49:35.644682 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898\": container with ID starting with 8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898 not found: ID does not exist" containerID="8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.644703 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898"} err="failed to get container status \"8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898\": rpc error: code = NotFound desc = could not find container \"8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898\": container with ID starting with 8ab24a9d66f32f93562227936cf665a7870c8c7f8d94f117cf7be619a8de3898 not found: ID does not exist" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.668514 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.668548 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/960e23d5-abd5-46aa-ba53-36c5e91b3a61-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.668562 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pzzf\" (UniqueName: \"kubernetes.io/projected/960e23d5-abd5-46aa-ba53-36c5e91b3a61-kube-api-access-2pzzf\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.899854 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gt8f6"] Oct 04 04:49:35 crc kubenswrapper[4992]: I1004 04:49:35.905966 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gt8f6"] Oct 04 04:49:36 crc kubenswrapper[4992]: I1004 04:49:36.333842 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" path="/var/lib/kubelet/pods/960e23d5-abd5-46aa-ba53-36c5e91b3a61/volumes" Oct 04 04:49:37 crc kubenswrapper[4992]: I1004 04:49:37.083020 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:37 crc kubenswrapper[4992]: I1004 04:49:37.083123 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:37 crc kubenswrapper[4992]: I1004 04:49:37.110428 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:37 crc kubenswrapper[4992]: I1004 04:49:37.610303 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xp6bx" Oct 04 04:49:37 crc kubenswrapper[4992]: I1004 04:49:37.969857 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg968"] Oct 04 04:49:37 crc kubenswrapper[4992]: I1004 04:49:37.970349 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zg968" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerName="registry-server" containerID="cri-o://437c8cb64151c5ac589541e201f39529ff462ff2cd55359cdc4eda0aa3880fe7" gracePeriod=2 Oct 04 04:49:38 crc kubenswrapper[4992]: I1004 04:49:38.332061 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-lb5zk" Oct 04 04:49:38 crc kubenswrapper[4992]: I1004 04:49:38.588324 4992 generic.go:334] "Generic (PLEG): container finished" podID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerID="437c8cb64151c5ac589541e201f39529ff462ff2cd55359cdc4eda0aa3880fe7" exitCode=0 Oct 04 04:49:38 crc kubenswrapper[4992]: I1004 04:49:38.588419 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg968" event={"ID":"82bcfc12-3f52-4014-a93e-98c05d64860f","Type":"ContainerDied","Data":"437c8cb64151c5ac589541e201f39529ff462ff2cd55359cdc4eda0aa3880fe7"} Oct 04 04:49:38 crc kubenswrapper[4992]: I1004 04:49:38.923818 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.117087 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-utilities\") pod \"82bcfc12-3f52-4014-a93e-98c05d64860f\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.117138 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4c77\" (UniqueName: \"kubernetes.io/projected/82bcfc12-3f52-4014-a93e-98c05d64860f-kube-api-access-j4c77\") pod \"82bcfc12-3f52-4014-a93e-98c05d64860f\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.117205 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-catalog-content\") pod \"82bcfc12-3f52-4014-a93e-98c05d64860f\" (UID: \"82bcfc12-3f52-4014-a93e-98c05d64860f\") " Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.118070 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-utilities" (OuterVolumeSpecName: "utilities") pod "82bcfc12-3f52-4014-a93e-98c05d64860f" (UID: "82bcfc12-3f52-4014-a93e-98c05d64860f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.122926 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82bcfc12-3f52-4014-a93e-98c05d64860f-kube-api-access-j4c77" (OuterVolumeSpecName: "kube-api-access-j4c77") pod "82bcfc12-3f52-4014-a93e-98c05d64860f" (UID: "82bcfc12-3f52-4014-a93e-98c05d64860f"). InnerVolumeSpecName "kube-api-access-j4c77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.132896 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82bcfc12-3f52-4014-a93e-98c05d64860f" (UID: "82bcfc12-3f52-4014-a93e-98c05d64860f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.218764 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.218799 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82bcfc12-3f52-4014-a93e-98c05d64860f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.218809 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4c77\" (UniqueName: \"kubernetes.io/projected/82bcfc12-3f52-4014-a93e-98c05d64860f-kube-api-access-j4c77\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.595870 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg968" event={"ID":"82bcfc12-3f52-4014-a93e-98c05d64860f","Type":"ContainerDied","Data":"298a2bbda5a68d06202b019cbdac6bfd86d88506c4d0bb4251a2ad2ddffa9407"} Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.595925 4992 scope.go:117] "RemoveContainer" containerID="437c8cb64151c5ac589541e201f39529ff462ff2cd55359cdc4eda0aa3880fe7" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.596041 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg968" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.611459 4992 scope.go:117] "RemoveContainer" containerID="f41ba888c3a839b392b81ddfa0f1e46f9103df1ea4785abba6c1d4fcfde68c2f" Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.625258 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg968"] Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.629533 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg968"] Oct 04 04:49:39 crc kubenswrapper[4992]: I1004 04:49:39.654730 4992 scope.go:117] "RemoveContainer" containerID="c41cce470a59e81cc8d21a264f22178981ee227f5e881c524d4aee4ff9f3e1b0" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.192489 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w"] Oct 04 04:49:40 crc kubenswrapper[4992]: E1004 04:49:40.192767 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerName="extract-content" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.192783 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerName="extract-content" Oct 04 04:49:40 crc kubenswrapper[4992]: E1004 04:49:40.192796 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerName="registry-server" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.192803 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerName="registry-server" Oct 04 04:49:40 crc kubenswrapper[4992]: E1004 04:49:40.192812 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerName="extract-utilities" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.192820 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerName="extract-utilities" Oct 04 04:49:40 crc kubenswrapper[4992]: E1004 04:49:40.192844 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerName="registry-server" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.192853 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerName="registry-server" Oct 04 04:49:40 crc kubenswrapper[4992]: E1004 04:49:40.192865 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerName="extract-utilities" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.192875 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerName="extract-utilities" Oct 04 04:49:40 crc kubenswrapper[4992]: E1004 04:49:40.192886 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerName="extract-content" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.192894 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerName="extract-content" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.193053 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="960e23d5-abd5-46aa-ba53-36c5e91b3a61" containerName="registry-server" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.193076 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" containerName="registry-server" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.194015 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.195933 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-gkwwf" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.200619 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w"] Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.231458 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt752\" (UniqueName: \"kubernetes.io/projected/c947400d-2e44-4d6d-b0c0-60814a3acdd6-kube-api-access-qt752\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.231823 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-bundle\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.231897 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-util\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.326229 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82bcfc12-3f52-4014-a93e-98c05d64860f" path="/var/lib/kubelet/pods/82bcfc12-3f52-4014-a93e-98c05d64860f/volumes" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.332673 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-util\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.332755 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt752\" (UniqueName: \"kubernetes.io/projected/c947400d-2e44-4d6d-b0c0-60814a3acdd6-kube-api-access-qt752\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.332799 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-bundle\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.333344 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-util\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.333376 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-bundle\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.355759 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt752\" (UniqueName: \"kubernetes.io/projected/c947400d-2e44-4d6d-b0c0-60814a3acdd6-kube-api-access-qt752\") pod \"53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.508864 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:40 crc kubenswrapper[4992]: I1004 04:49:40.973607 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w"] Oct 04 04:49:40 crc kubenswrapper[4992]: W1004 04:49:40.978070 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc947400d_2e44_4d6d_b0c0_60814a3acdd6.slice/crio-47cfba9230556aa7ae0dc297830c69852d19ad467472902f8e972dd5e264fac7 WatchSource:0}: Error finding container 47cfba9230556aa7ae0dc297830c69852d19ad467472902f8e972dd5e264fac7: Status 404 returned error can't find the container with id 47cfba9230556aa7ae0dc297830c69852d19ad467472902f8e972dd5e264fac7 Oct 04 04:49:41 crc kubenswrapper[4992]: I1004 04:49:41.612994 4992 generic.go:334] "Generic (PLEG): container finished" podID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerID="3c75cf64304e20221fea66f347462c34cf00f48fc178bbaa3d34958ac0b38bd6" exitCode=0 Oct 04 04:49:41 crc kubenswrapper[4992]: I1004 04:49:41.613035 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" event={"ID":"c947400d-2e44-4d6d-b0c0-60814a3acdd6","Type":"ContainerDied","Data":"3c75cf64304e20221fea66f347462c34cf00f48fc178bbaa3d34958ac0b38bd6"} Oct 04 04:49:41 crc kubenswrapper[4992]: I1004 04:49:41.613057 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" event={"ID":"c947400d-2e44-4d6d-b0c0-60814a3acdd6","Type":"ContainerStarted","Data":"47cfba9230556aa7ae0dc297830c69852d19ad467472902f8e972dd5e264fac7"} Oct 04 04:49:44 crc kubenswrapper[4992]: I1004 04:49:44.644215 4992 generic.go:334] "Generic (PLEG): container finished" podID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerID="fb3b2922465702a964e3838706083f26a7e71d82dcf15c409f9a41393c17df9d" exitCode=0 Oct 04 04:49:44 crc kubenswrapper[4992]: I1004 04:49:44.644305 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" event={"ID":"c947400d-2e44-4d6d-b0c0-60814a3acdd6","Type":"ContainerDied","Data":"fb3b2922465702a964e3838706083f26a7e71d82dcf15c409f9a41393c17df9d"} Oct 04 04:49:45 crc kubenswrapper[4992]: I1004 04:49:45.653504 4992 generic.go:334] "Generic (PLEG): container finished" podID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerID="587e4a8feff14a345d303461928b59b1d2b64863e31b1270041aab603d192daf" exitCode=0 Oct 04 04:49:45 crc kubenswrapper[4992]: I1004 04:49:45.653555 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" event={"ID":"c947400d-2e44-4d6d-b0c0-60814a3acdd6","Type":"ContainerDied","Data":"587e4a8feff14a345d303461928b59b1d2b64863e31b1270041aab603d192daf"} Oct 04 04:49:46 crc kubenswrapper[4992]: I1004 04:49:46.942591 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.129214 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-bundle\") pod \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.129611 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt752\" (UniqueName: \"kubernetes.io/projected/c947400d-2e44-4d6d-b0c0-60814a3acdd6-kube-api-access-qt752\") pod \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.129804 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-util\") pod \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\" (UID: \"c947400d-2e44-4d6d-b0c0-60814a3acdd6\") " Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.129931 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-bundle" (OuterVolumeSpecName: "bundle") pod "c947400d-2e44-4d6d-b0c0-60814a3acdd6" (UID: "c947400d-2e44-4d6d-b0c0-60814a3acdd6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.130093 4992 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.135124 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c947400d-2e44-4d6d-b0c0-60814a3acdd6-kube-api-access-qt752" (OuterVolumeSpecName: "kube-api-access-qt752") pod "c947400d-2e44-4d6d-b0c0-60814a3acdd6" (UID: "c947400d-2e44-4d6d-b0c0-60814a3acdd6"). InnerVolumeSpecName "kube-api-access-qt752". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.230859 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt752\" (UniqueName: \"kubernetes.io/projected/c947400d-2e44-4d6d-b0c0-60814a3acdd6-kube-api-access-qt752\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.667138 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" event={"ID":"c947400d-2e44-4d6d-b0c0-60814a3acdd6","Type":"ContainerDied","Data":"47cfba9230556aa7ae0dc297830c69852d19ad467472902f8e972dd5e264fac7"} Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.667179 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47cfba9230556aa7ae0dc297830c69852d19ad467472902f8e972dd5e264fac7" Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.667253 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w" Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.677766 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-util" (OuterVolumeSpecName: "util") pod "c947400d-2e44-4d6d-b0c0-60814a3acdd6" (UID: "c947400d-2e44-4d6d-b0c0-60814a3acdd6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:49:47 crc kubenswrapper[4992]: I1004 04:49:47.737034 4992 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c947400d-2e44-4d6d-b0c0-60814a3acdd6-util\") on node \"crc\" DevicePath \"\"" Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.894695 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.894997 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.973465 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq"] Oct 04 04:49:50 crc kubenswrapper[4992]: E1004 04:49:50.973780 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerName="pull" Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.973800 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerName="pull" Oct 04 04:49:50 crc kubenswrapper[4992]: E1004 04:49:50.973814 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerName="util" Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.973821 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerName="util" Oct 04 04:49:50 crc kubenswrapper[4992]: E1004 04:49:50.973829 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerName="extract" Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.973835 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerName="extract" Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.973958 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c947400d-2e44-4d6d-b0c0-60814a3acdd6" containerName="extract" Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.974700 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" Oct 04 04:49:50 crc kubenswrapper[4992]: I1004 04:49:50.976459 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-j6b22" Oct 04 04:49:51 crc kubenswrapper[4992]: I1004 04:49:51.005220 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq"] Oct 04 04:49:51 crc kubenswrapper[4992]: I1004 04:49:51.079266 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg4sf\" (UniqueName: \"kubernetes.io/projected/f3ae9a99-2aae-49d4-a881-8ff336de3c3f-kube-api-access-pg4sf\") pod \"openstack-operator-controller-operator-6dc5776fd9-k9tqq\" (UID: \"f3ae9a99-2aae-49d4-a881-8ff336de3c3f\") " pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" Oct 04 04:49:51 crc kubenswrapper[4992]: I1004 04:49:51.180733 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg4sf\" (UniqueName: \"kubernetes.io/projected/f3ae9a99-2aae-49d4-a881-8ff336de3c3f-kube-api-access-pg4sf\") pod \"openstack-operator-controller-operator-6dc5776fd9-k9tqq\" (UID: \"f3ae9a99-2aae-49d4-a881-8ff336de3c3f\") " pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" Oct 04 04:49:51 crc kubenswrapper[4992]: I1004 04:49:51.200222 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg4sf\" (UniqueName: \"kubernetes.io/projected/f3ae9a99-2aae-49d4-a881-8ff336de3c3f-kube-api-access-pg4sf\") pod \"openstack-operator-controller-operator-6dc5776fd9-k9tqq\" (UID: \"f3ae9a99-2aae-49d4-a881-8ff336de3c3f\") " pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" Oct 04 04:49:51 crc kubenswrapper[4992]: I1004 04:49:51.292714 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" Oct 04 04:49:51 crc kubenswrapper[4992]: I1004 04:49:51.714619 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq"] Oct 04 04:49:51 crc kubenswrapper[4992]: W1004 04:49:51.718408 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3ae9a99_2aae_49d4_a881_8ff336de3c3f.slice/crio-b7c850b94d920946a7658b4b1052ee72a0d2b2aac8e4866fc27a20473795c346 WatchSource:0}: Error finding container b7c850b94d920946a7658b4b1052ee72a0d2b2aac8e4866fc27a20473795c346: Status 404 returned error can't find the container with id b7c850b94d920946a7658b4b1052ee72a0d2b2aac8e4866fc27a20473795c346 Oct 04 04:49:52 crc kubenswrapper[4992]: I1004 04:49:52.699007 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" event={"ID":"f3ae9a99-2aae-49d4-a881-8ff336de3c3f","Type":"ContainerStarted","Data":"b7c850b94d920946a7658b4b1052ee72a0d2b2aac8e4866fc27a20473795c346"} Oct 04 04:50:02 crc kubenswrapper[4992]: I1004 04:50:02.772324 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" event={"ID":"f3ae9a99-2aae-49d4-a881-8ff336de3c3f","Type":"ContainerStarted","Data":"6e44eccdf157b57e1dc5536ca884c81700d12f477e4202b4a17593276a9adc8b"} Oct 04 04:50:07 crc kubenswrapper[4992]: I1004 04:50:07.805522 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" event={"ID":"f3ae9a99-2aae-49d4-a881-8ff336de3c3f","Type":"ContainerStarted","Data":"09722370a26c88582b82d9a7543a4dbea2fc995fcef75e3b9e62aa51df07d149"} Oct 04 04:50:07 crc kubenswrapper[4992]: I1004 04:50:07.805845 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" Oct 04 04:50:07 crc kubenswrapper[4992]: I1004 04:50:07.808988 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" Oct 04 04:50:07 crc kubenswrapper[4992]: I1004 04:50:07.846872 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6dc5776fd9-k9tqq" podStartSLOduration=1.932097528 podStartE2EDuration="17.846847689s" podCreationTimestamp="2025-10-04 04:49:50 +0000 UTC" firstStartedPulling="2025-10-04 04:49:51.720497128 +0000 UTC m=+1025.568172596" lastFinishedPulling="2025-10-04 04:50:07.635247299 +0000 UTC m=+1041.482922757" observedRunningTime="2025-10-04 04:50:07.843762766 +0000 UTC m=+1041.691438234" watchObservedRunningTime="2025-10-04 04:50:07.846847689 +0000 UTC m=+1041.694523167" Oct 04 04:50:20 crc kubenswrapper[4992]: I1004 04:50:20.894251 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:50:20 crc kubenswrapper[4992]: I1004 04:50:20.894653 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.725522 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.726869 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.733374 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-d7qt7" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.739454 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.749456 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.750579 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.753532 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-pj8pv" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.758523 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.759727 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.762048 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-85l66" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.783577 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.789812 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.799698 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.801067 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg2tw\" (UniqueName: \"kubernetes.io/projected/a6b05a48-2a86-4afc-87c4-e6006f45b812-kube-api-access-lg2tw\") pod \"barbican-operator-controller-manager-5f7c849b98-crk9q\" (UID: \"a6b05a48-2a86-4afc-87c4-e6006f45b812\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.808538 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.815779 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pw9lf" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.815982 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.866779 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.883691 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.895938 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-wg8cc" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.909036 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcnrl\" (UniqueName: \"kubernetes.io/projected/5c53b372-1912-4b4a-9a73-8b0619d4c9c8-kube-api-access-mcnrl\") pod \"glance-operator-controller-manager-5568b5d68-tctb4\" (UID: \"5c53b372-1912-4b4a-9a73-8b0619d4c9c8\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.912101 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg2tw\" (UniqueName: \"kubernetes.io/projected/a6b05a48-2a86-4afc-87c4-e6006f45b812-kube-api-access-lg2tw\") pod \"barbican-operator-controller-manager-5f7c849b98-crk9q\" (UID: \"a6b05a48-2a86-4afc-87c4-e6006f45b812\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.912230 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9bpg\" (UniqueName: \"kubernetes.io/projected/07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35-kube-api-access-v9bpg\") pod \"cinder-operator-controller-manager-7d4d4f8d-6gwsx\" (UID: \"07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.912356 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfrzq\" (UniqueName: \"kubernetes.io/projected/b185c908-306b-4a41-b5f4-94e6024b050b-kube-api-access-cfrzq\") pod \"designate-operator-controller-manager-75dfd9b554-scgjp\" (UID: \"b185c908-306b-4a41-b5f4-94e6024b050b\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.909528 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.921100 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.922664 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.926625 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.928020 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.929888 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.931701 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-5wfth" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.932551 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-k8scw" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.936661 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.942601 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.966325 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg2tw\" (UniqueName: \"kubernetes.io/projected/a6b05a48-2a86-4afc-87c4-e6006f45b812-kube-api-access-lg2tw\") pod \"barbican-operator-controller-manager-5f7c849b98-crk9q\" (UID: \"a6b05a48-2a86-4afc-87c4-e6006f45b812\") " pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.974428 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-576fh"] Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.977697 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.985974 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-zl8h2" Oct 04 04:50:22 crc kubenswrapper[4992]: I1004 04:50:22.997327 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-576fh"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:22.999982 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.003203 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.007738 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-k24lm" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.021776 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcnrl\" (UniqueName: \"kubernetes.io/projected/5c53b372-1912-4b4a-9a73-8b0619d4c9c8-kube-api-access-mcnrl\") pod \"glance-operator-controller-manager-5568b5d68-tctb4\" (UID: \"5c53b372-1912-4b4a-9a73-8b0619d4c9c8\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.021852 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9bpg\" (UniqueName: \"kubernetes.io/projected/07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35-kube-api-access-v9bpg\") pod \"cinder-operator-controller-manager-7d4d4f8d-6gwsx\" (UID: \"07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.021885 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqlcm\" (UniqueName: \"kubernetes.io/projected/7bb260b0-fb2e-4ad8-9ea9-198934b62e60-kube-api-access-wqlcm\") pod \"horizon-operator-controller-manager-54876c876f-r42pj\" (UID: \"7bb260b0-fb2e-4ad8-9ea9-198934b62e60\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.021951 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmftp\" (UniqueName: \"kubernetes.io/projected/fdba06a8-d608-4111-a067-aedd5f292557-kube-api-access-gmftp\") pod \"infra-operator-controller-manager-658588b8c9-czxxb\" (UID: \"fdba06a8-d608-4111-a067-aedd5f292557\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.022004 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfrzq\" (UniqueName: \"kubernetes.io/projected/b185c908-306b-4a41-b5f4-94e6024b050b-kube-api-access-cfrzq\") pod \"designate-operator-controller-manager-75dfd9b554-scgjp\" (UID: \"b185c908-306b-4a41-b5f4-94e6024b050b\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.022040 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fdba06a8-d608-4111-a067-aedd5f292557-cert\") pod \"infra-operator-controller-manager-658588b8c9-czxxb\" (UID: \"fdba06a8-d608-4111-a067-aedd5f292557\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.022100 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h894j\" (UniqueName: \"kubernetes.io/projected/bd358fd0-4f19-4420-9dee-7d8a2daafb83-kube-api-access-h894j\") pod \"heat-operator-controller-manager-8f58bc9db-59d5d\" (UID: \"bd358fd0-4f19-4420-9dee-7d8a2daafb83\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.029291 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.033471 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.034448 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.037199 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.042703 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.043841 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.050442 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-sk4v2" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.050563 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-pkbg7" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.056098 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.060798 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcnrl\" (UniqueName: \"kubernetes.io/projected/5c53b372-1912-4b4a-9a73-8b0619d4c9c8-kube-api-access-mcnrl\") pod \"glance-operator-controller-manager-5568b5d68-tctb4\" (UID: \"5c53b372-1912-4b4a-9a73-8b0619d4c9c8\") " pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.066312 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9bpg\" (UniqueName: \"kubernetes.io/projected/07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35-kube-api-access-v9bpg\") pod \"cinder-operator-controller-manager-7d4d4f8d-6gwsx\" (UID: \"07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.070563 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.071704 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.077419 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-vz4pr" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.084125 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfrzq\" (UniqueName: \"kubernetes.io/projected/b185c908-306b-4a41-b5f4-94e6024b050b-kube-api-access-cfrzq\") pod \"designate-operator-controller-manager-75dfd9b554-scgjp\" (UID: \"b185c908-306b-4a41-b5f4-94e6024b050b\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.084212 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.091334 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.096902 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.102876 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.104708 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.114702 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-t8c47" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.115044 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.125048 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fdba06a8-d608-4111-a067-aedd5f292557-cert\") pod \"infra-operator-controller-manager-658588b8c9-czxxb\" (UID: \"fdba06a8-d608-4111-a067-aedd5f292557\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.125102 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfnfj\" (UniqueName: \"kubernetes.io/projected/4185f955-0390-47f4-88d3-18f5c7f9c4e5-kube-api-access-gfnfj\") pod \"keystone-operator-controller-manager-655d88ccb9-5j2nq\" (UID: \"4185f955-0390-47f4-88d3-18f5c7f9c4e5\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.125137 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpvbq\" (UniqueName: \"kubernetes.io/projected/44ed3dc1-bf8a-4c86-90c2-71490a4ab882-kube-api-access-fpvbq\") pod \"ironic-operator-controller-manager-699b87f775-576fh\" (UID: \"44ed3dc1-bf8a-4c86-90c2-71490a4ab882\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.125166 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h894j\" (UniqueName: \"kubernetes.io/projected/bd358fd0-4f19-4420-9dee-7d8a2daafb83-kube-api-access-h894j\") pod \"heat-operator-controller-manager-8f58bc9db-59d5d\" (UID: \"bd358fd0-4f19-4420-9dee-7d8a2daafb83\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.125213 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6h6m\" (UniqueName: \"kubernetes.io/projected/411b554a-a210-4025-be0b-ba8d0a6e912b-kube-api-access-g6h6m\") pod \"manila-operator-controller-manager-65d89cfd9f-4vlxm\" (UID: \"411b554a-a210-4025-be0b-ba8d0a6e912b\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.125242 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdwmq\" (UniqueName: \"kubernetes.io/projected/7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668-kube-api-access-rdwmq\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-t52mq\" (UID: \"7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.125294 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqlcm\" (UniqueName: \"kubernetes.io/projected/7bb260b0-fb2e-4ad8-9ea9-198934b62e60-kube-api-access-wqlcm\") pod \"horizon-operator-controller-manager-54876c876f-r42pj\" (UID: \"7bb260b0-fb2e-4ad8-9ea9-198934b62e60\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.125330 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmftp\" (UniqueName: \"kubernetes.io/projected/fdba06a8-d608-4111-a067-aedd5f292557-kube-api-access-gmftp\") pod \"infra-operator-controller-manager-658588b8c9-czxxb\" (UID: \"fdba06a8-d608-4111-a067-aedd5f292557\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:23 crc kubenswrapper[4992]: E1004 04:50:23.125340 4992 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 04 04:50:23 crc kubenswrapper[4992]: E1004 04:50:23.125420 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fdba06a8-d608-4111-a067-aedd5f292557-cert podName:fdba06a8-d608-4111-a067-aedd5f292557 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:23.625399574 +0000 UTC m=+1057.473075132 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fdba06a8-d608-4111-a067-aedd5f292557-cert") pod "infra-operator-controller-manager-658588b8c9-czxxb" (UID: "fdba06a8-d608-4111-a067-aedd5f292557") : secret "infra-operator-webhook-server-cert" not found Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.138265 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.139521 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.148915 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.161819 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-crz5k" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.170454 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.173327 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqlcm\" (UniqueName: \"kubernetes.io/projected/7bb260b0-fb2e-4ad8-9ea9-198934b62e60-kube-api-access-wqlcm\") pod \"horizon-operator-controller-manager-54876c876f-r42pj\" (UID: \"7bb260b0-fb2e-4ad8-9ea9-198934b62e60\") " pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.182226 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmftp\" (UniqueName: \"kubernetes.io/projected/fdba06a8-d608-4111-a067-aedd5f292557-kube-api-access-gmftp\") pod \"infra-operator-controller-manager-658588b8c9-czxxb\" (UID: \"fdba06a8-d608-4111-a067-aedd5f292557\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.183382 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.187879 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.195752 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-vpzvf" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.205625 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.206668 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.207570 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h894j\" (UniqueName: \"kubernetes.io/projected/bd358fd0-4f19-4420-9dee-7d8a2daafb83-kube-api-access-h894j\") pod \"heat-operator-controller-manager-8f58bc9db-59d5d\" (UID: \"bd358fd0-4f19-4420-9dee-7d8a2daafb83\") " pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.211055 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.212738 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rxvc6" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.212937 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.225643 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.227100 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfnfj\" (UniqueName: \"kubernetes.io/projected/4185f955-0390-47f4-88d3-18f5c7f9c4e5-kube-api-access-gfnfj\") pod \"keystone-operator-controller-manager-655d88ccb9-5j2nq\" (UID: \"4185f955-0390-47f4-88d3-18f5c7f9c4e5\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.227129 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpvbq\" (UniqueName: \"kubernetes.io/projected/44ed3dc1-bf8a-4c86-90c2-71490a4ab882-kube-api-access-fpvbq\") pod \"ironic-operator-controller-manager-699b87f775-576fh\" (UID: \"44ed3dc1-bf8a-4c86-90c2-71490a4ab882\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.227159 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7rz4\" (UniqueName: \"kubernetes.io/projected/94a4d61f-bf8a-436a-a267-f8b014d9418e-kube-api-access-v7rz4\") pod \"nova-operator-controller-manager-7c7fc454ff-mj9np\" (UID: \"94a4d61f-bf8a-436a-a267-f8b014d9418e\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.227188 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6h6m\" (UniqueName: \"kubernetes.io/projected/411b554a-a210-4025-be0b-ba8d0a6e912b-kube-api-access-g6h6m\") pod \"manila-operator-controller-manager-65d89cfd9f-4vlxm\" (UID: \"411b554a-a210-4025-be0b-ba8d0a6e912b\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.227206 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmq\" (UniqueName: \"kubernetes.io/projected/7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668-kube-api-access-rdwmq\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-t52mq\" (UID: \"7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.227235 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94qbb\" (UniqueName: \"kubernetes.io/projected/63cd2106-f93c-49af-88f3-a1598ab76e8b-kube-api-access-94qbb\") pod \"octavia-operator-controller-manager-7468f855d8-cnzlq\" (UID: \"63cd2106-f93c-49af-88f3-a1598ab76e8b\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.227255 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kw75\" (UniqueName: \"kubernetes.io/projected/49e9ad0f-706f-4aa5-8e5d-347e29a408cc-kube-api-access-2kw75\") pod \"neutron-operator-controller-manager-8d984cc4d-nbbkf\" (UID: \"49e9ad0f-706f-4aa5-8e5d-347e29a408cc\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.227302 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.229425 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.239761 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-8hkt7" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.239938 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.240984 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.245809 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-gwn5l" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.261310 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.273907 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmq\" (UniqueName: \"kubernetes.io/projected/7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668-kube-api-access-rdwmq\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-t52mq\" (UID: \"7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.277966 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.279099 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpvbq\" (UniqueName: \"kubernetes.io/projected/44ed3dc1-bf8a-4c86-90c2-71490a4ab882-kube-api-access-fpvbq\") pod \"ironic-operator-controller-manager-699b87f775-576fh\" (UID: \"44ed3dc1-bf8a-4c86-90c2-71490a4ab882\") " pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.279554 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfnfj\" (UniqueName: \"kubernetes.io/projected/4185f955-0390-47f4-88d3-18f5c7f9c4e5-kube-api-access-gfnfj\") pod \"keystone-operator-controller-manager-655d88ccb9-5j2nq\" (UID: \"4185f955-0390-47f4-88d3-18f5c7f9c4e5\") " pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.279947 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.285102 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-fbssm" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.288492 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.297851 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6h6m\" (UniqueName: \"kubernetes.io/projected/411b554a-a210-4025-be0b-ba8d0a6e912b-kube-api-access-g6h6m\") pod \"manila-operator-controller-manager-65d89cfd9f-4vlxm\" (UID: \"411b554a-a210-4025-be0b-ba8d0a6e912b\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.305052 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.306073 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.336340 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.349890 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.350613 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.354439 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p6jv\" (UniqueName: \"kubernetes.io/projected/d0d758eb-b8a5-4508-bdc2-8a58158d864f-kube-api-access-9p6jv\") pod \"swift-operator-controller-manager-6859f9b676-dc9gg\" (UID: \"d0d758eb-b8a5-4508-bdc2-8a58158d864f\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.354546 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9hh7\" (UniqueName: \"kubernetes.io/projected/75dc0bae-78b9-45f6-b8fd-f9ee9f307ace-kube-api-access-k9hh7\") pod \"ovn-operator-controller-manager-579449c7d5-85jrh\" (UID: \"75dc0bae-78b9-45f6-b8fd-f9ee9f307ace\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.354681 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2\" (UID: \"70c5935b-fd73-4f67-928f-178131a1c7c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.355188 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7rz4\" (UniqueName: \"kubernetes.io/projected/94a4d61f-bf8a-436a-a267-f8b014d9418e-kube-api-access-v7rz4\") pod \"nova-operator-controller-manager-7c7fc454ff-mj9np\" (UID: \"94a4d61f-bf8a-436a-a267-f8b014d9418e\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.355345 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwnns\" (UniqueName: \"kubernetes.io/projected/70c5935b-fd73-4f67-928f-178131a1c7c5-kube-api-access-hwnns\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2\" (UID: \"70c5935b-fd73-4f67-928f-178131a1c7c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.356568 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl4dj\" (UniqueName: \"kubernetes.io/projected/463db7ad-b3a2-469e-9a74-cb170fb54b0d-kube-api-access-hl4dj\") pod \"placement-operator-controller-manager-54689d9f88-vmkh5\" (UID: \"463db7ad-b3a2-469e-9a74-cb170fb54b0d\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.357661 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94qbb\" (UniqueName: \"kubernetes.io/projected/63cd2106-f93c-49af-88f3-a1598ab76e8b-kube-api-access-94qbb\") pod \"octavia-operator-controller-manager-7468f855d8-cnzlq\" (UID: \"63cd2106-f93c-49af-88f3-a1598ab76e8b\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.358582 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kw75\" (UniqueName: \"kubernetes.io/projected/49e9ad0f-706f-4aa5-8e5d-347e29a408cc-kube-api-access-2kw75\") pod \"neutron-operator-controller-manager-8d984cc4d-nbbkf\" (UID: \"49e9ad0f-706f-4aa5-8e5d-347e29a408cc\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.363192 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-d4cn8" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.366782 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.397114 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kw75\" (UniqueName: \"kubernetes.io/projected/49e9ad0f-706f-4aa5-8e5d-347e29a408cc-kube-api-access-2kw75\") pod \"neutron-operator-controller-manager-8d984cc4d-nbbkf\" (UID: \"49e9ad0f-706f-4aa5-8e5d-347e29a408cc\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.397152 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.399113 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94qbb\" (UniqueName: \"kubernetes.io/projected/63cd2106-f93c-49af-88f3-a1598ab76e8b-kube-api-access-94qbb\") pod \"octavia-operator-controller-manager-7468f855d8-cnzlq\" (UID: \"63cd2106-f93c-49af-88f3-a1598ab76e8b\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.399295 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7rz4\" (UniqueName: \"kubernetes.io/projected/94a4d61f-bf8a-436a-a267-f8b014d9418e-kube-api-access-v7rz4\") pod \"nova-operator-controller-manager-7c7fc454ff-mj9np\" (UID: \"94a4d61f-bf8a-436a-a267-f8b014d9418e\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.399826 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.407506 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.423766 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.457521 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.459788 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p6jv\" (UniqueName: \"kubernetes.io/projected/d0d758eb-b8a5-4508-bdc2-8a58158d864f-kube-api-access-9p6jv\") pod \"swift-operator-controller-manager-6859f9b676-dc9gg\" (UID: \"d0d758eb-b8a5-4508-bdc2-8a58158d864f\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.459852 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9hh7\" (UniqueName: \"kubernetes.io/projected/75dc0bae-78b9-45f6-b8fd-f9ee9f307ace-kube-api-access-k9hh7\") pod \"ovn-operator-controller-manager-579449c7d5-85jrh\" (UID: \"75dc0bae-78b9-45f6-b8fd-f9ee9f307ace\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.459926 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2\" (UID: \"70c5935b-fd73-4f67-928f-178131a1c7c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:23 crc kubenswrapper[4992]: E1004 04:50:23.460754 4992 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 04:50:23 crc kubenswrapper[4992]: E1004 04:50:23.460824 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert podName:70c5935b-fd73-4f67-928f-178131a1c7c5 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:23.960806365 +0000 UTC m=+1057.808481833 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" (UID: "70c5935b-fd73-4f67-928f-178131a1c7c5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.462853 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.468841 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m2fv\" (UniqueName: \"kubernetes.io/projected/892ab60b-1964-4943-9238-eedb8a72b3e4-kube-api-access-8m2fv\") pod \"telemetry-operator-controller-manager-5d4d74dd89-phbgj\" (UID: \"892ab60b-1964-4943-9238-eedb8a72b3e4\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.468980 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwnns\" (UniqueName: \"kubernetes.io/projected/70c5935b-fd73-4f67-928f-178131a1c7c5-kube-api-access-hwnns\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2\" (UID: \"70c5935b-fd73-4f67-928f-178131a1c7c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.469024 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl4dj\" (UniqueName: \"kubernetes.io/projected/463db7ad-b3a2-469e-9a74-cb170fb54b0d-kube-api-access-hl4dj\") pod \"placement-operator-controller-manager-54689d9f88-vmkh5\" (UID: \"463db7ad-b3a2-469e-9a74-cb170fb54b0d\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.469075 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n5cb\" (UniqueName: \"kubernetes.io/projected/ca9f0c2e-15e3-404f-83b6-a9a438ca88c9-kube-api-access-6n5cb\") pod \"test-operator-controller-manager-5cd5cb47d7-kf7nd\" (UID: \"ca9f0c2e-15e3-404f-83b6-a9a438ca88c9\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.474507 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.476099 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.479443 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-b2gkr" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.499471 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9hh7\" (UniqueName: \"kubernetes.io/projected/75dc0bae-78b9-45f6-b8fd-f9ee9f307ace-kube-api-access-k9hh7\") pod \"ovn-operator-controller-manager-579449c7d5-85jrh\" (UID: \"75dc0bae-78b9-45f6-b8fd-f9ee9f307ace\") " pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.511107 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl4dj\" (UniqueName: \"kubernetes.io/projected/463db7ad-b3a2-469e-9a74-cb170fb54b0d-kube-api-access-hl4dj\") pod \"placement-operator-controller-manager-54689d9f88-vmkh5\" (UID: \"463db7ad-b3a2-469e-9a74-cb170fb54b0d\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.512588 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwnns\" (UniqueName: \"kubernetes.io/projected/70c5935b-fd73-4f67-928f-178131a1c7c5-kube-api-access-hwnns\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2\" (UID: \"70c5935b-fd73-4f67-928f-178131a1c7c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.518657 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.525118 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p6jv\" (UniqueName: \"kubernetes.io/projected/d0d758eb-b8a5-4508-bdc2-8a58158d864f-kube-api-access-9p6jv\") pod \"swift-operator-controller-manager-6859f9b676-dc9gg\" (UID: \"d0d758eb-b8a5-4508-bdc2-8a58158d864f\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.528264 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.540531 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.596253 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m2fv\" (UniqueName: \"kubernetes.io/projected/892ab60b-1964-4943-9238-eedb8a72b3e4-kube-api-access-8m2fv\") pod \"telemetry-operator-controller-manager-5d4d74dd89-phbgj\" (UID: \"892ab60b-1964-4943-9238-eedb8a72b3e4\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.596405 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n5cb\" (UniqueName: \"kubernetes.io/projected/ca9f0c2e-15e3-404f-83b6-a9a438ca88c9-kube-api-access-6n5cb\") pod \"test-operator-controller-manager-5cd5cb47d7-kf7nd\" (UID: \"ca9f0c2e-15e3-404f-83b6-a9a438ca88c9\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.596558 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9q7q\" (UniqueName: \"kubernetes.io/projected/cbcb9dfe-fcdc-4adf-b65c-346aded84b3f-kube-api-access-v9q7q\") pod \"watcher-operator-controller-manager-6cbc6dd547-mdvtr\" (UID: \"cbcb9dfe-fcdc-4adf-b65c-346aded84b3f\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.619035 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.625736 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.625843 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.627554 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n5cb\" (UniqueName: \"kubernetes.io/projected/ca9f0c2e-15e3-404f-83b6-a9a438ca88c9-kube-api-access-6n5cb\") pod \"test-operator-controller-manager-5cd5cb47d7-kf7nd\" (UID: \"ca9f0c2e-15e3-404f-83b6-a9a438ca88c9\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.627838 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.631638 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.632443 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m2fv\" (UniqueName: \"kubernetes.io/projected/892ab60b-1964-4943-9238-eedb8a72b3e4-kube-api-access-8m2fv\") pod \"telemetry-operator-controller-manager-5d4d74dd89-phbgj\" (UID: \"892ab60b-1964-4943-9238-eedb8a72b3e4\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.632516 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.632845 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-gthnx" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.633216 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.635925 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-k465t" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.640168 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.697947 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9q7q\" (UniqueName: \"kubernetes.io/projected/cbcb9dfe-fcdc-4adf-b65c-346aded84b3f-kube-api-access-v9q7q\") pod \"watcher-operator-controller-manager-6cbc6dd547-mdvtr\" (UID: \"cbcb9dfe-fcdc-4adf-b65c-346aded84b3f\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.698028 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fdba06a8-d608-4111-a067-aedd5f292557-cert\") pod \"infra-operator-controller-manager-658588b8c9-czxxb\" (UID: \"fdba06a8-d608-4111-a067-aedd5f292557\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.713187 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fdba06a8-d608-4111-a067-aedd5f292557-cert\") pod \"infra-operator-controller-manager-658588b8c9-czxxb\" (UID: \"fdba06a8-d608-4111-a067-aedd5f292557\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.716880 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9q7q\" (UniqueName: \"kubernetes.io/projected/cbcb9dfe-fcdc-4adf-b65c-346aded84b3f-kube-api-access-v9q7q\") pod \"watcher-operator-controller-manager-6cbc6dd547-mdvtr\" (UID: \"cbcb9dfe-fcdc-4adf-b65c-346aded84b3f\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.752734 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.801624 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qc5x\" (UniqueName: \"kubernetes.io/projected/ef4005a7-625d-49a8-a37a-25058eb5de15-kube-api-access-9qc5x\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt\" (UID: \"ef4005a7-625d-49a8-a37a-25058eb5de15\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.801714 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfrjs\" (UniqueName: \"kubernetes.io/projected/9736ec36-6280-4353-9e71-f64c8e6b5756-kube-api-access-vfrjs\") pod \"openstack-operator-controller-manager-6c6db87b77-kftlz\" (UID: \"9736ec36-6280-4353-9e71-f64c8e6b5756\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.801765 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-kftlz\" (UID: \"9736ec36-6280-4353-9e71-f64c8e6b5756\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.809304 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.866061 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.908583 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfrjs\" (UniqueName: \"kubernetes.io/projected/9736ec36-6280-4353-9e71-f64c8e6b5756-kube-api-access-vfrjs\") pod \"openstack-operator-controller-manager-6c6db87b77-kftlz\" (UID: \"9736ec36-6280-4353-9e71-f64c8e6b5756\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.908695 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-kftlz\" (UID: \"9736ec36-6280-4353-9e71-f64c8e6b5756\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.908860 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qc5x\" (UniqueName: \"kubernetes.io/projected/ef4005a7-625d-49a8-a37a-25058eb5de15-kube-api-access-9qc5x\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt\" (UID: \"ef4005a7-625d-49a8-a37a-25058eb5de15\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" Oct 04 04:50:23 crc kubenswrapper[4992]: E1004 04:50:23.909027 4992 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 04:50:23 crc kubenswrapper[4992]: E1004 04:50:23.909082 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert podName:9736ec36-6280-4353-9e71-f64c8e6b5756 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:24.409063185 +0000 UTC m=+1058.256738653 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert") pod "openstack-operator-controller-manager-6c6db87b77-kftlz" (UID: "9736ec36-6280-4353-9e71-f64c8e6b5756") : secret "webhook-server-cert" not found Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.917803 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.919821 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q"] Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.924659 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.929137 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.939138 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfrjs\" (UniqueName: \"kubernetes.io/projected/9736ec36-6280-4353-9e71-f64c8e6b5756-kube-api-access-vfrjs\") pod \"openstack-operator-controller-manager-6c6db87b77-kftlz\" (UID: \"9736ec36-6280-4353-9e71-f64c8e6b5756\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.941092 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qc5x\" (UniqueName: \"kubernetes.io/projected/ef4005a7-625d-49a8-a37a-25058eb5de15-kube-api-access-9qc5x\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt\" (UID: \"ef4005a7-625d-49a8-a37a-25058eb5de15\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" Oct 04 04:50:23 crc kubenswrapper[4992]: I1004 04:50:23.988162 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.015813 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2\" (UID: \"70c5935b-fd73-4f67-928f-178131a1c7c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.016057 4992 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.016117 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert podName:70c5935b-fd73-4f67-928f-178131a1c7c5 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:25.016100268 +0000 UTC m=+1058.863775736 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" (UID: "70c5935b-fd73-4f67-928f-178131a1c7c5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.250901 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.259703 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-699b87f775-576fh"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.276479 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx"] Oct 04 04:50:24 crc kubenswrapper[4992]: W1004 04:50:24.315910 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07db2e1d_eb98_4ea9_bcda_a1ce07ecbd35.slice/crio-4b4aae9ac9be87f7faf1c59e0c873320d05031d1f082950effde2300cc1f02a6 WatchSource:0}: Error finding container 4b4aae9ac9be87f7faf1c59e0c873320d05031d1f082950effde2300cc1f02a6: Status 404 returned error can't find the container with id 4b4aae9ac9be87f7faf1c59e0c873320d05031d1f082950effde2300cc1f02a6 Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.352663 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.379079 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.435026 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-kftlz\" (UID: \"9736ec36-6280-4353-9e71-f64c8e6b5756\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.436295 4992 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.436628 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert podName:9736ec36-6280-4353-9e71-f64c8e6b5756 nodeName:}" failed. No retries permitted until 2025-10-04 04:50:25.436603498 +0000 UTC m=+1059.284278966 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert") pod "openstack-operator-controller-manager-6c6db87b77-kftlz" (UID: "9736ec36-6280-4353-9e71-f64c8e6b5756") : secret "webhook-server-cert" not found Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.544140 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj"] Oct 04 04:50:24 crc kubenswrapper[4992]: W1004 04:50:24.641101 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75dc0bae_78b9_45f6_b8fd_f9ee9f307ace.slice/crio-4940c0bd09f893a205e9d63146f6bd38df70573c9ccc3529adcf9c24e65ed134 WatchSource:0}: Error finding container 4940c0bd09f893a205e9d63146f6bd38df70573c9ccc3529adcf9c24e65ed134: Status 404 returned error can't find the container with id 4940c0bd09f893a205e9d63146f6bd38df70573c9ccc3529adcf9c24e65ed134 Oct 04 04:50:24 crc kubenswrapper[4992]: W1004 04:50:24.642450 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63cd2106_f93c_49af_88f3_a1598ab76e8b.slice/crio-07cd8b92125b82f8002e735ea177449f0801bd40493654f64aa1be79463b49f2 WatchSource:0}: Error finding container 07cd8b92125b82f8002e735ea177449f0801bd40493654f64aa1be79463b49f2: Status 404 returned error can't find the container with id 07cd8b92125b82f8002e735ea177449f0801bd40493654f64aa1be79463b49f2 Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.646148 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.651978 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.661457 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm"] Oct 04 04:50:24 crc kubenswrapper[4992]: W1004 04:50:24.665736 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod463db7ad_b3a2_469e_9a74_cb170fb54b0d.slice/crio-79dd1eb8910f8bf1f67630da11e03615a8dd89824d5d571c630dae07f9a933c7 WatchSource:0}: Error finding container 79dd1eb8910f8bf1f67630da11e03615a8dd89824d5d571c630dae07f9a933c7: Status 404 returned error can't find the container with id 79dd1eb8910f8bf1f67630da11e03615a8dd89824d5d571c630dae07f9a933c7 Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.669181 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.834314 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.848572 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.856152 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.861539 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np"] Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.869147 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:585796b996a5b6d7ad68f0cb420bf4f2ee38c9f16f194e3111c162ce91ea8a7b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cfrzq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-75dfd9b554-scgjp_openstack-operators(b185c908-306b-4a41-b5f4-94e6024b050b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.869267 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd"] Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.869942 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gmftp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-658588b8c9-czxxb_openstack-operators(fdba06a8-d608-4111-a067-aedd5f292557): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.869277 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2kw75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-8d984cc4d-nbbkf_openstack-operators(49e9ad0f-706f-4aa5-8e5d-347e29a408cc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:50:24 crc kubenswrapper[4992]: W1004 04:50:24.881018 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca9f0c2e_15e3_404f_83b6_a9a438ca88c9.slice/crio-aae40afb79c752f05ab10c7fea2e36a04d0474ca71c457daf89af2917247f6f0 WatchSource:0}: Error finding container aae40afb79c752f05ab10c7fea2e36a04d0474ca71c457daf89af2917247f6f0: Status 404 returned error can't find the container with id aae40afb79c752f05ab10c7fea2e36a04d0474ca71c457daf89af2917247f6f0 Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.881032 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8m2fv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d4d74dd89-phbgj_openstack-operators(892ab60b-1964-4943-9238-eedb8a72b3e4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.885571 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp"] Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.888223 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9p6jv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6859f9b676-dc9gg_openstack-operators(d0d758eb-b8a5-4508-bdc2-8a58158d864f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.888434 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v9q7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6cbc6dd547-mdvtr_openstack-operators(cbcb9dfe-fcdc-4adf-b65c-346aded84b3f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:50:24 crc kubenswrapper[4992]: E1004 04:50:24.888445 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6n5cb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd5cb47d7-kf7nd_openstack-operators(ca9f0c2e-15e3-404f-83b6-a9a438ca88c9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.892955 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.895917 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.903399 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.905771 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt"] Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.948672 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" event={"ID":"63cd2106-f93c-49af-88f3-a1598ab76e8b","Type":"ContainerStarted","Data":"07cd8b92125b82f8002e735ea177449f0801bd40493654f64aa1be79463b49f2"} Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.953974 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" event={"ID":"411b554a-a210-4025-be0b-ba8d0a6e912b","Type":"ContainerStarted","Data":"48ea12bd9cebe277e45cbb54bfdc6a6ed094e293ae4bfef6a3083fb540630688"} Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.955259 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" event={"ID":"b185c908-306b-4a41-b5f4-94e6024b050b","Type":"ContainerStarted","Data":"5916bcf0f60be32227ba0038281587d12791a4dd23b99d8f1aa9f129d54702d8"} Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.958564 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" event={"ID":"94a4d61f-bf8a-436a-a267-f8b014d9418e","Type":"ContainerStarted","Data":"4eb834df1627cabe8bc4e1a4160bc8b8c0e9539c828c5793297c6037f683e582"} Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.960470 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" event={"ID":"07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35","Type":"ContainerStarted","Data":"4b4aae9ac9be87f7faf1c59e0c873320d05031d1f082950effde2300cc1f02a6"} Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.961348 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" event={"ID":"7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668","Type":"ContainerStarted","Data":"a17ed2d5044c409006cf511c12c6c16be3424f4ce5feefadbc3867f9a7636340"} Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.986619 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" event={"ID":"44ed3dc1-bf8a-4c86-90c2-71490a4ab882","Type":"ContainerStarted","Data":"cd240c0a559607ddb0aadbd927b6dc51cb0a8413dd36b0d6f293680bd50779d7"} Oct 04 04:50:24 crc kubenswrapper[4992]: I1004 04:50:24.990594 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" event={"ID":"5c53b372-1912-4b4a-9a73-8b0619d4c9c8","Type":"ContainerStarted","Data":"9282d952dae9343293354761c35186e67626c4bba6122480049bfc07af1d6f5b"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.000563 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" event={"ID":"fdba06a8-d608-4111-a067-aedd5f292557","Type":"ContainerStarted","Data":"2faee6698a17c3242b356b5f9da9daec14d39df1cafeb11467e87f2c300a9117"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.004218 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" event={"ID":"cbcb9dfe-fcdc-4adf-b65c-346aded84b3f","Type":"ContainerStarted","Data":"d69747646a3be2ad206e1ac106011e746599ad8b9e267357c59b6c8bea2a38b5"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.006939 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" event={"ID":"bd358fd0-4f19-4420-9dee-7d8a2daafb83","Type":"ContainerStarted","Data":"c3527bdb71dd7db558021ab31470f0d8df5734332983c6d810065374f8026319"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.008843 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" event={"ID":"4185f955-0390-47f4-88d3-18f5c7f9c4e5","Type":"ContainerStarted","Data":"f967b55576279aff3ded2e477b4ea17d6131ad1493df381e86a1e7b5c8737740"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.011861 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" event={"ID":"49e9ad0f-706f-4aa5-8e5d-347e29a408cc","Type":"ContainerStarted","Data":"9ebfcbcb126954e6576cc951d0fbbb0f209e610226420367515688c366f20426"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.017348 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" event={"ID":"ca9f0c2e-15e3-404f-83b6-a9a438ca88c9","Type":"ContainerStarted","Data":"aae40afb79c752f05ab10c7fea2e36a04d0474ca71c457daf89af2917247f6f0"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.029833 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" event={"ID":"7bb260b0-fb2e-4ad8-9ea9-198934b62e60","Type":"ContainerStarted","Data":"be6d6b7cfc3252dad9107de1ffba46863f5902a37a0089434c33e86535b3f8bb"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.031510 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" event={"ID":"892ab60b-1964-4943-9238-eedb8a72b3e4","Type":"ContainerStarted","Data":"e1709b6285a77d07812f03c7396f28becace7555d35dfb6c5432a81773598572"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.032996 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" event={"ID":"75dc0bae-78b9-45f6-b8fd-f9ee9f307ace","Type":"ContainerStarted","Data":"4940c0bd09f893a205e9d63146f6bd38df70573c9ccc3529adcf9c24e65ed134"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.035167 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" event={"ID":"a6b05a48-2a86-4afc-87c4-e6006f45b812","Type":"ContainerStarted","Data":"3b297a4dcb9514e4216d4087a19a3bfca286211ca168c4e4b333c66e72858f82"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.036600 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" event={"ID":"d0d758eb-b8a5-4508-bdc2-8a58158d864f","Type":"ContainerStarted","Data":"29004e739044f79ab5fcae75aaaf268844f5cb334d68e111de35f6c1b7fb5fa3"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.039874 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" event={"ID":"463db7ad-b3a2-469e-9a74-cb170fb54b0d","Type":"ContainerStarted","Data":"79dd1eb8910f8bf1f67630da11e03615a8dd89824d5d571c630dae07f9a933c7"} Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.053816 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2\" (UID: \"70c5935b-fd73-4f67-928f-178131a1c7c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.065958 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70c5935b-fd73-4f67-928f-178131a1c7c5-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2\" (UID: \"70c5935b-fd73-4f67-928f-178131a1c7c5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.277032 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:25 crc kubenswrapper[4992]: E1004 04:50:25.447781 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" podUID="cbcb9dfe-fcdc-4adf-b65c-346aded84b3f" Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.469406 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-kftlz\" (UID: \"9736ec36-6280-4353-9e71-f64c8e6b5756\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.473686 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9736ec36-6280-4353-9e71-f64c8e6b5756-cert\") pod \"openstack-operator-controller-manager-6c6db87b77-kftlz\" (UID: \"9736ec36-6280-4353-9e71-f64c8e6b5756\") " pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:25 crc kubenswrapper[4992]: E1004 04:50:25.486690 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" podUID="b185c908-306b-4a41-b5f4-94e6024b050b" Oct 04 04:50:25 crc kubenswrapper[4992]: E1004 04:50:25.491681 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" podUID="fdba06a8-d608-4111-a067-aedd5f292557" Oct 04 04:50:25 crc kubenswrapper[4992]: E1004 04:50:25.491986 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" podUID="49e9ad0f-706f-4aa5-8e5d-347e29a408cc" Oct 04 04:50:25 crc kubenswrapper[4992]: E1004 04:50:25.496533 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" podUID="d0d758eb-b8a5-4508-bdc2-8a58158d864f" Oct 04 04:50:25 crc kubenswrapper[4992]: E1004 04:50:25.499652 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" podUID="892ab60b-1964-4943-9238-eedb8a72b3e4" Oct 04 04:50:25 crc kubenswrapper[4992]: E1004 04:50:25.524952 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" podUID="ca9f0c2e-15e3-404f-83b6-a9a438ca88c9" Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.710101 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2"] Oct 04 04:50:25 crc kubenswrapper[4992]: I1004 04:50:25.756909 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.056093 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" event={"ID":"70c5935b-fd73-4f67-928f-178131a1c7c5","Type":"ContainerStarted","Data":"6928b2bcc0d78d12a63c7ba120904b7869e02298cb3c1f09a351b223faae66a2"} Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.066744 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" event={"ID":"ca9f0c2e-15e3-404f-83b6-a9a438ca88c9","Type":"ContainerStarted","Data":"ddda27f26749c7ebda3cfa2ab4352d3c7b51e59c09322376c8bc1644b3e790b7"} Oct 04 04:50:26 crc kubenswrapper[4992]: E1004 04:50:26.070610 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" podUID="ca9f0c2e-15e3-404f-83b6-a9a438ca88c9" Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.093567 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" event={"ID":"fdba06a8-d608-4111-a067-aedd5f292557","Type":"ContainerStarted","Data":"64d01a95140cca6c848e4131fa18a18c353f5c881a572b74e965bf3c686ebf6f"} Oct 04 04:50:26 crc kubenswrapper[4992]: E1004 04:50:26.095641 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" podUID="fdba06a8-d608-4111-a067-aedd5f292557" Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.101395 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" event={"ID":"b185c908-306b-4a41-b5f4-94e6024b050b","Type":"ContainerStarted","Data":"f7e0615cec463c44804ea97921ba805bfa9bd420051291e11b5979683202fffc"} Oct 04 04:50:26 crc kubenswrapper[4992]: E1004 04:50:26.103437 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:585796b996a5b6d7ad68f0cb420bf4f2ee38c9f16f194e3111c162ce91ea8a7b\\\"\"" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" podUID="b185c908-306b-4a41-b5f4-94e6024b050b" Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.107464 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" event={"ID":"cbcb9dfe-fcdc-4adf-b65c-346aded84b3f","Type":"ContainerStarted","Data":"164519da735422df5df14fceef2dca0402189d95b38abcd6444b9d27fbbe4999"} Oct 04 04:50:26 crc kubenswrapper[4992]: E1004 04:50:26.115898 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" podUID="cbcb9dfe-fcdc-4adf-b65c-346aded84b3f" Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.165507 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" event={"ID":"ef4005a7-625d-49a8-a37a-25058eb5de15","Type":"ContainerStarted","Data":"8f987e27237c544547a2cbb92d65ea99cc2b6f9379c2f5fdc4dfacee698c6eef"} Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.206738 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" event={"ID":"892ab60b-1964-4943-9238-eedb8a72b3e4","Type":"ContainerStarted","Data":"43950e8b8724321b9c664cf512b882fe2c01b8eba04b55f53fd62a6871ce9635"} Oct 04 04:50:26 crc kubenswrapper[4992]: E1004 04:50:26.212238 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" podUID="892ab60b-1964-4943-9238-eedb8a72b3e4" Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.214472 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" event={"ID":"d0d758eb-b8a5-4508-bdc2-8a58158d864f","Type":"ContainerStarted","Data":"8127bed3c887fd9143a2f27b7b013db2da4e17b24f3e0225f3bf65b85b251e58"} Oct 04 04:50:26 crc kubenswrapper[4992]: E1004 04:50:26.216720 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" podUID="d0d758eb-b8a5-4508-bdc2-8a58158d864f" Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.216935 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" event={"ID":"49e9ad0f-706f-4aa5-8e5d-347e29a408cc","Type":"ContainerStarted","Data":"955147963a17a76a4c06a1811a4356c54311de334cab60ac98f2fb8086ea052f"} Oct 04 04:50:26 crc kubenswrapper[4992]: E1004 04:50:26.233637 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" podUID="49e9ad0f-706f-4aa5-8e5d-347e29a408cc" Oct 04 04:50:26 crc kubenswrapper[4992]: I1004 04:50:26.455142 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz"] Oct 04 04:50:27 crc kubenswrapper[4992]: I1004 04:50:27.255574 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" event={"ID":"9736ec36-6280-4353-9e71-f64c8e6b5756","Type":"ContainerStarted","Data":"6e7ff828429c6ec7a3abf603f10c1cd6599a4f3857de69490cc10528a787c9e8"} Oct 04 04:50:27 crc kubenswrapper[4992]: I1004 04:50:27.255925 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" event={"ID":"9736ec36-6280-4353-9e71-f64c8e6b5756","Type":"ContainerStarted","Data":"ec9ccf18437ba29388fd36fa913b624c0b128571a30056c41ee5aa89853a09e7"} Oct 04 04:50:27 crc kubenswrapper[4992]: I1004 04:50:27.255937 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" event={"ID":"9736ec36-6280-4353-9e71-f64c8e6b5756","Type":"ContainerStarted","Data":"1950b867839b841c10a8ff3d7ad7d5ede2da377dec4f6d9311612728c69d1fcb"} Oct 04 04:50:27 crc kubenswrapper[4992]: E1004 04:50:27.258500 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" podUID="cbcb9dfe-fcdc-4adf-b65c-346aded84b3f" Oct 04 04:50:27 crc kubenswrapper[4992]: E1004 04:50:27.258510 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:585796b996a5b6d7ad68f0cb420bf4f2ee38c9f16f194e3111c162ce91ea8a7b\\\"\"" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" podUID="b185c908-306b-4a41-b5f4-94e6024b050b" Oct 04 04:50:27 crc kubenswrapper[4992]: E1004 04:50:27.258508 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:b6cef68bfaacdf992a9fa1a6b03a848a48c18cbb6ed12d95561b4b37d858b99f\\\"\"" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" podUID="fdba06a8-d608-4111-a067-aedd5f292557" Oct 04 04:50:27 crc kubenswrapper[4992]: E1004 04:50:27.258541 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:dfd044635f9df9ed1d249387fa622177db35cdc72475e1c570617b8d17c64862\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" podUID="49e9ad0f-706f-4aa5-8e5d-347e29a408cc" Oct 04 04:50:27 crc kubenswrapper[4992]: E1004 04:50:27.258554 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:637bb7b9ac308bc1e323391a3593b824f688090a856c83385814c17a571b1eed\\\"\"" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" podUID="d0d758eb-b8a5-4508-bdc2-8a58158d864f" Oct 04 04:50:27 crc kubenswrapper[4992]: E1004 04:50:27.258581 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" podUID="ca9f0c2e-15e3-404f-83b6-a9a438ca88c9" Oct 04 04:50:27 crc kubenswrapper[4992]: E1004 04:50:27.258597 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" podUID="892ab60b-1964-4943-9238-eedb8a72b3e4" Oct 04 04:50:27 crc kubenswrapper[4992]: I1004 04:50:27.327125 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" podStartSLOduration=4.327109351 podStartE2EDuration="4.327109351s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:50:27.326608518 +0000 UTC m=+1061.174283986" watchObservedRunningTime="2025-10-04 04:50:27.327109351 +0000 UTC m=+1061.174784819" Oct 04 04:50:28 crc kubenswrapper[4992]: I1004 04:50:28.289412 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:35 crc kubenswrapper[4992]: I1004 04:50:35.764959 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6c6db87b77-kftlz" Oct 04 04:50:38 crc kubenswrapper[4992]: E1004 04:50:38.750968 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Oct 04 04:50:38 crc kubenswrapper[4992]: E1004 04:50:38.751802 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9qc5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt_openstack-operators(ef4005a7-625d-49a8-a37a-25058eb5de15): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:50:38 crc kubenswrapper[4992]: E1004 04:50:38.753042 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" podUID="ef4005a7-625d-49a8-a37a-25058eb5de15" Oct 04 04:50:39 crc kubenswrapper[4992]: E1004 04:50:39.268136 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182" Oct 04 04:50:39 crc kubenswrapper[4992]: E1004 04:50:39.268528 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-94qbb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7468f855d8-cnzlq_openstack-operators(63cd2106-f93c-49af-88f3-a1598ab76e8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:50:39 crc kubenswrapper[4992]: E1004 04:50:39.362325 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" podUID="ef4005a7-625d-49a8-a37a-25058eb5de15" Oct 04 04:50:39 crc kubenswrapper[4992]: E1004 04:50:39.809095 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254" Oct 04 04:50:39 crc kubenswrapper[4992]: E1004 04:50:39.809348 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k9hh7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-579449c7d5-85jrh_openstack-operators(75dc0bae-78b9-45f6-b8fd-f9ee9f307ace): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:50:40 crc kubenswrapper[4992]: E1004 04:50:40.348044 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:daed26a1dce4b8221159a011cbc5905ef63e9887fdca2118e4bcc61f88b5fb0a" Oct 04 04:50:40 crc kubenswrapper[4992]: E1004 04:50:40.348701 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:daed26a1dce4b8221159a011cbc5905ef63e9887fdca2118e4bcc61f88b5fb0a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gfnfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-655d88ccb9-5j2nq_openstack-operators(4185f955-0390-47f4-88d3-18f5c7f9c4e5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:50:41 crc kubenswrapper[4992]: E1004 04:50:41.778922 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" podUID="4185f955-0390-47f4-88d3-18f5c7f9c4e5" Oct 04 04:50:41 crc kubenswrapper[4992]: E1004 04:50:41.781323 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" podUID="75dc0bae-78b9-45f6-b8fd-f9ee9f307ace" Oct 04 04:50:41 crc kubenswrapper[4992]: E1004 04:50:41.793437 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" podUID="63cd2106-f93c-49af-88f3-a1598ab76e8b" Oct 04 04:50:42 crc kubenswrapper[4992]: I1004 04:50:42.382542 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" event={"ID":"4185f955-0390-47f4-88d3-18f5c7f9c4e5","Type":"ContainerStarted","Data":"30e446ad702dc7fd5564dbfa7813ec0ac4af1dca59149b9601e4c5443a21918d"} Oct 04 04:50:42 crc kubenswrapper[4992]: I1004 04:50:42.385453 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" event={"ID":"75dc0bae-78b9-45f6-b8fd-f9ee9f307ace","Type":"ContainerStarted","Data":"0801db3d03125c4ebaaf9129b738e9ec505bdaa19fe7d3d648d24a73cebeb3c3"} Oct 04 04:50:42 crc kubenswrapper[4992]: I1004 04:50:42.389019 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" event={"ID":"63cd2106-f93c-49af-88f3-a1598ab76e8b","Type":"ContainerStarted","Data":"ad52cf79cc13dfb41de8ac50a667d1117a9929951c4759aaac0ac290b92a4183"} Oct 04 04:50:42 crc kubenswrapper[4992]: I1004 04:50:42.391399 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" event={"ID":"411b554a-a210-4025-be0b-ba8d0a6e912b","Type":"ContainerStarted","Data":"b722aa1c8367bd8401d816214a5c50e3b8cba2ebd48e2bed206ecfff96a30ebf"} Oct 04 04:50:42 crc kubenswrapper[4992]: E1004 04:50:42.407656 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" podUID="75dc0bae-78b9-45f6-b8fd-f9ee9f307ace" Oct 04 04:50:42 crc kubenswrapper[4992]: E1004 04:50:42.407902 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" podUID="63cd2106-f93c-49af-88f3-a1598ab76e8b" Oct 04 04:50:42 crc kubenswrapper[4992]: E1004 04:50:42.408042 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:daed26a1dce4b8221159a011cbc5905ef63e9887fdca2118e4bcc61f88b5fb0a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" podUID="4185f955-0390-47f4-88d3-18f5c7f9c4e5" Oct 04 04:50:43 crc kubenswrapper[4992]: I1004 04:50:43.399824 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" event={"ID":"a6b05a48-2a86-4afc-87c4-e6006f45b812","Type":"ContainerStarted","Data":"69ba65dc7760aa5b92ab329b5b51b494e738fd575637bab90e983ba9540a9f5c"} Oct 04 04:50:43 crc kubenswrapper[4992]: E1004 04:50:43.404151 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" podUID="75dc0bae-78b9-45f6-b8fd-f9ee9f307ace" Oct 04 04:50:43 crc kubenswrapper[4992]: E1004 04:50:43.404902 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" podUID="63cd2106-f93c-49af-88f3-a1598ab76e8b" Oct 04 04:50:43 crc kubenswrapper[4992]: E1004 04:50:43.405775 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:daed26a1dce4b8221159a011cbc5905ef63e9887fdca2118e4bcc61f88b5fb0a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" podUID="4185f955-0390-47f4-88d3-18f5c7f9c4e5" Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.409322 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" event={"ID":"07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35","Type":"ContainerStarted","Data":"8aa3bc7dc9a412974104c7c5b0fc28b9163b6908bf48fbc3ff577336f29f664a"} Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.412490 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" event={"ID":"44ed3dc1-bf8a-4c86-90c2-71490a4ab882","Type":"ContainerStarted","Data":"18d8133f6cf424db110720bc30224360f1a5f8552cb7fbb516206a423875345a"} Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.415071 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" event={"ID":"b185c908-306b-4a41-b5f4-94e6024b050b","Type":"ContainerStarted","Data":"e64703f7cfa5316bac0563ac862d4ab322f5dbde298e1e8a22b46c7ed6193357"} Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.415297 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.427547 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" event={"ID":"5c53b372-1912-4b4a-9a73-8b0619d4c9c8","Type":"ContainerStarted","Data":"377fac310021cff2d121cc2f90331a3360c0b7755db7cc3a670ec42e827bc8db"} Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.439750 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" podStartSLOduration=5.770048019 podStartE2EDuration="22.439727678s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.869021965 +0000 UTC m=+1058.716697433" lastFinishedPulling="2025-10-04 04:50:41.538701624 +0000 UTC m=+1075.386377092" observedRunningTime="2025-10-04 04:50:44.433562173 +0000 UTC m=+1078.281237661" watchObservedRunningTime="2025-10-04 04:50:44.439727678 +0000 UTC m=+1078.287403146" Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.446913 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" event={"ID":"7bb260b0-fb2e-4ad8-9ea9-198934b62e60","Type":"ContainerStarted","Data":"5b08ebe481f98d993fd2669ce3e0f46227bc6f2a67f32e0b4c989b1326a5e553"} Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.449277 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" event={"ID":"463db7ad-b3a2-469e-9a74-cb170fb54b0d","Type":"ContainerStarted","Data":"7c6a76ed23e09d931f03613e8fc61d44d3f1119a466ae8637b94a4675f10efca"} Oct 04 04:50:44 crc kubenswrapper[4992]: I1004 04:50:44.451078 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" event={"ID":"bd358fd0-4f19-4420-9dee-7d8a2daafb83","Type":"ContainerStarted","Data":"7e6d213ec84aa5987c25b3afcc0abe0b29a245fe6cebf4db56602ee4d704ea2b"} Oct 04 04:50:45 crc kubenswrapper[4992]: I1004 04:50:45.468494 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" event={"ID":"fdba06a8-d608-4111-a067-aedd5f292557","Type":"ContainerStarted","Data":"d00da23a2c0b646dc51d767d3842baf8660e51f3a4b4584602ebd2989394e7bb"} Oct 04 04:50:45 crc kubenswrapper[4992]: I1004 04:50:45.469002 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:45 crc kubenswrapper[4992]: I1004 04:50:45.470446 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" event={"ID":"94a4d61f-bf8a-436a-a267-f8b014d9418e","Type":"ContainerStarted","Data":"b74398d825a27d5d78c1e1896cdec9a29879441c34dd174044ee885da781e9c3"} Oct 04 04:50:45 crc kubenswrapper[4992]: I1004 04:50:45.473235 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" event={"ID":"7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668","Type":"ContainerStarted","Data":"eef4b56374de0eaafae549ce10b22e0f53c3bb2c31f5ef31e54019c1d0d15569"} Oct 04 04:50:45 crc kubenswrapper[4992]: I1004 04:50:45.485031 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" podStartSLOduration=6.816526916 podStartE2EDuration="23.484991182s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.869871218 +0000 UTC m=+1058.717546686" lastFinishedPulling="2025-10-04 04:50:41.538335484 +0000 UTC m=+1075.386010952" observedRunningTime="2025-10-04 04:50:45.481351695 +0000 UTC m=+1079.329027173" watchObservedRunningTime="2025-10-04 04:50:45.484991182 +0000 UTC m=+1079.332666650" Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.481186 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" event={"ID":"411b554a-a210-4025-be0b-ba8d0a6e912b","Type":"ContainerStarted","Data":"50307e4fcaf003fb472568b1a1848aa5e90976d77d27c673f7a7290859561aa0"} Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.481466 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.483001 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.484847 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" event={"ID":"ca9f0c2e-15e3-404f-83b6-a9a438ca88c9","Type":"ContainerStarted","Data":"206964748c7aa143d089788f819bbb0539634c42fc618f87de6396bd620b7594"} Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.487114 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" event={"ID":"463db7ad-b3a2-469e-9a74-cb170fb54b0d","Type":"ContainerStarted","Data":"7311f0e1bf222deb26b55d5534072cab3900e23bce8110613b47db9f7de1873f"} Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.487253 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.489235 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" event={"ID":"70c5935b-fd73-4f67-928f-178131a1c7c5","Type":"ContainerStarted","Data":"1619f032a701c0c92e230ee87035ec325f1a1867740010214452f36fa6d9560e"} Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.497828 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-4vlxm" podStartSLOduration=7.650750954 podStartE2EDuration="24.497808761s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.651652521 +0000 UTC m=+1058.499327989" lastFinishedPulling="2025-10-04 04:50:41.498710328 +0000 UTC m=+1075.346385796" observedRunningTime="2025-10-04 04:50:46.495665864 +0000 UTC m=+1080.343341332" watchObservedRunningTime="2025-10-04 04:50:46.497808761 +0000 UTC m=+1080.345484229" Oct 04 04:50:46 crc kubenswrapper[4992]: I1004 04:50:46.551016 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" podStartSLOduration=6.724704125 podStartE2EDuration="23.550998059s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.669434365 +0000 UTC m=+1058.517109833" lastFinishedPulling="2025-10-04 04:50:41.495728299 +0000 UTC m=+1075.343403767" observedRunningTime="2025-10-04 04:50:46.54577705 +0000 UTC m=+1080.393452518" watchObservedRunningTime="2025-10-04 04:50:46.550998059 +0000 UTC m=+1080.398673527" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.499677 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" event={"ID":"44ed3dc1-bf8a-4c86-90c2-71490a4ab882","Type":"ContainerStarted","Data":"836497ff3928a73d7670bcdb389879850a6c5d3bae58fd91e075b3ebe6ed916a"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.500023 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.501702 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" event={"ID":"d0d758eb-b8a5-4508-bdc2-8a58158d864f","Type":"ContainerStarted","Data":"5fac4009171b2664ed8affb0e094b71b6873945256488c56359347c8c2e30067"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.501911 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.503833 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" event={"ID":"49e9ad0f-706f-4aa5-8e5d-347e29a408cc","Type":"ContainerStarted","Data":"f34cf642899ae50d1561d7dbe413272f2e5a4437b49d43ebd956d5bbc031ff12"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.504931 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.508467 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" event={"ID":"70c5935b-fd73-4f67-928f-178131a1c7c5","Type":"ContainerStarted","Data":"b4100805d3bdcc156dd6e13d68f1cbfbdc13348cd5e2d88ce1eba01e71142c71"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.509027 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.510644 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" event={"ID":"7bb260b0-fb2e-4ad8-9ea9-198934b62e60","Type":"ContainerStarted","Data":"a9b0c74860e1a73c28ed182b561c42fee5422776b621ebf8f55574579c76351a"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.511126 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.517259 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" event={"ID":"bd358fd0-4f19-4420-9dee-7d8a2daafb83","Type":"ContainerStarted","Data":"4b112bd9ca1bee338d5cc941f444a22d8f95baf14a419cab5f78c5c223f97a72"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.517971 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.518651 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" podStartSLOduration=8.319747037 podStartE2EDuration="25.518627553s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.296145864 +0000 UTC m=+1058.143821332" lastFinishedPulling="2025-10-04 04:50:41.49502638 +0000 UTC m=+1075.342701848" observedRunningTime="2025-10-04 04:50:47.517940294 +0000 UTC m=+1081.365615762" watchObservedRunningTime="2025-10-04 04:50:47.518627553 +0000 UTC m=+1081.366303041" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.522912 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" event={"ID":"07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35","Type":"ContainerStarted","Data":"c8b56beb3fd5f5d2855e29421eaa41759685a231552a788bd0af5e48bf5fd867"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.523126 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.525180 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" event={"ID":"a6b05a48-2a86-4afc-87c4-e6006f45b812","Type":"ContainerStarted","Data":"dd2c9fecc1db3190fdc72ded8780e45c5a4c62fd6694655f75e132b606ae5d65"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.526088 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.529917 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.533092 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" event={"ID":"7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668","Type":"ContainerStarted","Data":"b07899988a6ff9cad2bf099492028da913de3be9375e5e3d40000102332219e5"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.533338 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.535506 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" event={"ID":"cbcb9dfe-fcdc-4adf-b65c-346aded84b3f","Type":"ContainerStarted","Data":"8cdebdb6e112755708e4e6856c4ac01f1c4b2782e1eb9679bd11e3681277a546"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.535717 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.543273 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" event={"ID":"94a4d61f-bf8a-436a-a267-f8b014d9418e","Type":"ContainerStarted","Data":"1a48f25e58b9191e84ebb10395e72611b07d9dc18d9a6cce0e801bc904e8c818"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.543411 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.545587 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" event={"ID":"892ab60b-1964-4943-9238-eedb8a72b3e4","Type":"ContainerStarted","Data":"49f5bc11bd2ec3a031a0df1a60e8d5d52414c36c26597cb7e3f519915f50e216"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.545801 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.553298 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" event={"ID":"5c53b372-1912-4b4a-9a73-8b0619d4c9c8","Type":"ContainerStarted","Data":"b8a088bd94fbd3f1a8a52fd899199a5ba7db203a35db91ab4b3e5447d99f96fc"} Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.554591 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" podStartSLOduration=8.805083231 podStartE2EDuration="24.554573951s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:25.749978408 +0000 UTC m=+1059.597653876" lastFinishedPulling="2025-10-04 04:50:41.499469128 +0000 UTC m=+1075.347144596" observedRunningTime="2025-10-04 04:50:47.551942981 +0000 UTC m=+1081.399618459" watchObservedRunningTime="2025-10-04 04:50:47.554573951 +0000 UTC m=+1081.402249419" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.575667 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" podStartSLOduration=8.468551204 podStartE2EDuration="25.575643163s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.391935667 +0000 UTC m=+1058.239611135" lastFinishedPulling="2025-10-04 04:50:41.499027626 +0000 UTC m=+1075.346703094" observedRunningTime="2025-10-04 04:50:47.572782406 +0000 UTC m=+1081.420457904" watchObservedRunningTime="2025-10-04 04:50:47.575643163 +0000 UTC m=+1081.423318631" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.626839 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" podStartSLOduration=5.780514906 podStartE2EDuration="24.626821207s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.888090414 +0000 UTC m=+1058.735765882" lastFinishedPulling="2025-10-04 04:50:43.734396715 +0000 UTC m=+1077.582072183" observedRunningTime="2025-10-04 04:50:47.624217118 +0000 UTC m=+1081.471892586" watchObservedRunningTime="2025-10-04 04:50:47.626821207 +0000 UTC m=+1081.474496675" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.628390 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" podStartSLOduration=8.678958683 podStartE2EDuration="25.628381399s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.547421542 +0000 UTC m=+1058.395097010" lastFinishedPulling="2025-10-04 04:50:41.496844258 +0000 UTC m=+1075.344519726" observedRunningTime="2025-10-04 04:50:47.595581024 +0000 UTC m=+1081.443256492" watchObservedRunningTime="2025-10-04 04:50:47.628381399 +0000 UTC m=+1081.476056857" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.674348 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" podStartSLOduration=4.383395023 podStartE2EDuration="25.674325343s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.86920772 +0000 UTC m=+1058.716883188" lastFinishedPulling="2025-10-04 04:50:46.16013803 +0000 UTC m=+1080.007813508" observedRunningTime="2025-10-04 04:50:47.652625785 +0000 UTC m=+1081.500301273" watchObservedRunningTime="2025-10-04 04:50:47.674325343 +0000 UTC m=+1081.522000811" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.686018 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" podStartSLOduration=8.526388536 podStartE2EDuration="25.686001795s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.339939101 +0000 UTC m=+1058.187614569" lastFinishedPulling="2025-10-04 04:50:41.49955236 +0000 UTC m=+1075.347227828" observedRunningTime="2025-10-04 04:50:47.683829457 +0000 UTC m=+1081.531504925" watchObservedRunningTime="2025-10-04 04:50:47.686001795 +0000 UTC m=+1081.533677253" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.686200 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" podStartSLOduration=3.423835692 podStartE2EDuration="24.68619348s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.88833306 +0000 UTC m=+1058.736008528" lastFinishedPulling="2025-10-04 04:50:46.150690848 +0000 UTC m=+1079.998366316" observedRunningTime="2025-10-04 04:50:47.671819356 +0000 UTC m=+1081.519494834" watchObservedRunningTime="2025-10-04 04:50:47.68619348 +0000 UTC m=+1081.533868948" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.708910 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" podStartSLOduration=8.510332018 podStartE2EDuration="25.708892305s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.296517614 +0000 UTC m=+1058.144193072" lastFinishedPulling="2025-10-04 04:50:41.495077891 +0000 UTC m=+1075.342753359" observedRunningTime="2025-10-04 04:50:47.703802979 +0000 UTC m=+1081.551478447" watchObservedRunningTime="2025-10-04 04:50:47.708892305 +0000 UTC m=+1081.556567773" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.727083 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" podStartSLOduration=4.828717662 podStartE2EDuration="24.727065149s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.880914512 +0000 UTC m=+1058.728589970" lastFinishedPulling="2025-10-04 04:50:44.779261989 +0000 UTC m=+1078.626937457" observedRunningTime="2025-10-04 04:50:47.721981564 +0000 UTC m=+1081.569657042" watchObservedRunningTime="2025-10-04 04:50:47.727065149 +0000 UTC m=+1081.574740617" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.740526 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" podStartSLOduration=8.636396068 podStartE2EDuration="25.740506238s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.392603925 +0000 UTC m=+1058.240279393" lastFinishedPulling="2025-10-04 04:50:41.496714095 +0000 UTC m=+1075.344389563" observedRunningTime="2025-10-04 04:50:47.735510894 +0000 UTC m=+1081.583186372" watchObservedRunningTime="2025-10-04 04:50:47.740506238 +0000 UTC m=+1081.588181706" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.757933 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5f7c849b98-crk9q" podStartSLOduration=8.878604835 podStartE2EDuration="25.757915772s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.016949191 +0000 UTC m=+1057.864624659" lastFinishedPulling="2025-10-04 04:50:40.896260128 +0000 UTC m=+1074.743935596" observedRunningTime="2025-10-04 04:50:47.756406251 +0000 UTC m=+1081.604081719" watchObservedRunningTime="2025-10-04 04:50:47.757915772 +0000 UTC m=+1081.605591230" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.775632 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" podStartSLOduration=5.917327962 podStartE2EDuration="24.775615523s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.888296509 +0000 UTC m=+1058.735971977" lastFinishedPulling="2025-10-04 04:50:43.74658407 +0000 UTC m=+1077.594259538" observedRunningTime="2025-10-04 04:50:47.774056952 +0000 UTC m=+1081.621732420" watchObservedRunningTime="2025-10-04 04:50:47.775615523 +0000 UTC m=+1081.623290991" Oct 04 04:50:47 crc kubenswrapper[4992]: I1004 04:50:47.796091 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" podStartSLOduration=9.150614985 podStartE2EDuration="25.796071209s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.855151275 +0000 UTC m=+1058.702826743" lastFinishedPulling="2025-10-04 04:50:41.500607499 +0000 UTC m=+1075.348282967" observedRunningTime="2025-10-04 04:50:47.794668661 +0000 UTC m=+1081.642344139" watchObservedRunningTime="2025-10-04 04:50:47.796071209 +0000 UTC m=+1081.643746677" Oct 04 04:50:48 crc kubenswrapper[4992]: I1004 04:50:48.563170 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" Oct 04 04:50:48 crc kubenswrapper[4992]: I1004 04:50:48.564233 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-8f58bc9db-59d5d" Oct 04 04:50:48 crc kubenswrapper[4992]: I1004 04:50:48.565117 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-699b87f775-576fh" Oct 04 04:50:48 crc kubenswrapper[4992]: I1004 04:50:48.565194 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-6gwsx" Oct 04 04:50:48 crc kubenswrapper[4992]: I1004 04:50:48.565618 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5568b5d68-tctb4" Oct 04 04:50:48 crc kubenswrapper[4992]: I1004 04:50:48.565764 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54876c876f-r42pj" Oct 04 04:50:50 crc kubenswrapper[4992]: I1004 04:50:50.894973 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:50:50 crc kubenswrapper[4992]: I1004 04:50:50.895310 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:50:50 crc kubenswrapper[4992]: I1004 04:50:50.895388 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:50:50 crc kubenswrapper[4992]: I1004 04:50:50.896053 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb88432e3860828bfc70223a5d4820bc8d7427ffafdb37bf5816fc3fffccf08d"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:50:50 crc kubenswrapper[4992]: I1004 04:50:50.896115 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://bb88432e3860828bfc70223a5d4820bc8d7427ffafdb37bf5816fc3fffccf08d" gracePeriod=600 Oct 04 04:50:51 crc kubenswrapper[4992]: I1004 04:50:51.581854 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="bb88432e3860828bfc70223a5d4820bc8d7427ffafdb37bf5816fc3fffccf08d" exitCode=0 Oct 04 04:50:51 crc kubenswrapper[4992]: I1004 04:50:51.581937 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"bb88432e3860828bfc70223a5d4820bc8d7427ffafdb37bf5816fc3fffccf08d"} Oct 04 04:50:51 crc kubenswrapper[4992]: I1004 04:50:51.582265 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"5c66b07d0127863fb32b4550ae384badc1dd16f486eaff1e71f5b9fa6a2b2449"} Oct 04 04:50:51 crc kubenswrapper[4992]: I1004 04:50:51.582288 4992 scope.go:117] "RemoveContainer" containerID="391734b7432560d7fffa75f76956def8a70e770707e3c21215c4ea365ca34204" Oct 04 04:50:52 crc kubenswrapper[4992]: I1004 04:50:52.595561 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" event={"ID":"ef4005a7-625d-49a8-a37a-25058eb5de15","Type":"ContainerStarted","Data":"b84de4c569456ba26c71397d3ddda46cfc05052b538f36dedd73a82acb56f071"} Oct 04 04:50:52 crc kubenswrapper[4992]: I1004 04:50:52.610575 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt" podStartSLOduration=2.410686025 podStartE2EDuration="29.610554101s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.955752107 +0000 UTC m=+1058.803427575" lastFinishedPulling="2025-10-04 04:50:52.155620173 +0000 UTC m=+1086.003295651" observedRunningTime="2025-10-04 04:50:52.608765283 +0000 UTC m=+1086.456440801" watchObservedRunningTime="2025-10-04 04:50:52.610554101 +0000 UTC m=+1086.458229579" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.315964 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-t52mq" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.400986 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-scgjp" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.405219 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-mj9np" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.544781 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-vmkh5" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.630002 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-nbbkf" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.811780 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-dc9gg" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.868729 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-phbgj" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.918855 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.920764 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-kf7nd" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.929859 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-czxxb" Oct 04 04:50:53 crc kubenswrapper[4992]: I1004 04:50:53.931118 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-mdvtr" Oct 04 04:50:55 crc kubenswrapper[4992]: I1004 04:50:55.283106 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2" Oct 04 04:50:55 crc kubenswrapper[4992]: I1004 04:50:55.619090 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" event={"ID":"63cd2106-f93c-49af-88f3-a1598ab76e8b","Type":"ContainerStarted","Data":"a7e8e3bc10504aff5d148c86d9eab8e403ef18f6a610693e033e97c399a2c720"} Oct 04 04:50:55 crc kubenswrapper[4992]: I1004 04:50:55.620002 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" Oct 04 04:50:55 crc kubenswrapper[4992]: I1004 04:50:55.636830 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" podStartSLOduration=2.57820031 podStartE2EDuration="32.636810811s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.651209209 +0000 UTC m=+1058.498884677" lastFinishedPulling="2025-10-04 04:50:54.70981971 +0000 UTC m=+1088.557495178" observedRunningTime="2025-10-04 04:50:55.632762753 +0000 UTC m=+1089.480438221" watchObservedRunningTime="2025-10-04 04:50:55.636810811 +0000 UTC m=+1089.484486279" Oct 04 04:50:57 crc kubenswrapper[4992]: I1004 04:50:57.643647 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" event={"ID":"75dc0bae-78b9-45f6-b8fd-f9ee9f307ace","Type":"ContainerStarted","Data":"c24f4c171eab510a2f32fd484234ad6205f862728b6c1f04c33c05a37d469001"} Oct 04 04:50:57 crc kubenswrapper[4992]: I1004 04:50:57.644099 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" Oct 04 04:50:57 crc kubenswrapper[4992]: I1004 04:50:57.674832 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" podStartSLOduration=2.584147828 podStartE2EDuration="34.674811338s" podCreationTimestamp="2025-10-04 04:50:23 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.642754343 +0000 UTC m=+1058.490429811" lastFinishedPulling="2025-10-04 04:50:56.733417853 +0000 UTC m=+1090.581093321" observedRunningTime="2025-10-04 04:50:57.667394911 +0000 UTC m=+1091.515070379" watchObservedRunningTime="2025-10-04 04:50:57.674811338 +0000 UTC m=+1091.522486806" Oct 04 04:50:59 crc kubenswrapper[4992]: I1004 04:50:59.660701 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" event={"ID":"4185f955-0390-47f4-88d3-18f5c7f9c4e5","Type":"ContainerStarted","Data":"9c4ff566afcbbd4664f49e5a06a98a17d0a6153b47f085b978ac6df7b051d39f"} Oct 04 04:50:59 crc kubenswrapper[4992]: I1004 04:50:59.661238 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" Oct 04 04:50:59 crc kubenswrapper[4992]: I1004 04:50:59.681833 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" podStartSLOduration=3.830729472 podStartE2EDuration="37.681811998s" podCreationTimestamp="2025-10-04 04:50:22 +0000 UTC" firstStartedPulling="2025-10-04 04:50:24.854675523 +0000 UTC m=+1058.702350991" lastFinishedPulling="2025-10-04 04:50:58.705758049 +0000 UTC m=+1092.553433517" observedRunningTime="2025-10-04 04:50:59.676123827 +0000 UTC m=+1093.523799305" watchObservedRunningTime="2025-10-04 04:50:59.681811998 +0000 UTC m=+1093.529487466" Oct 04 04:51:03 crc kubenswrapper[4992]: I1004 04:51:03.429059 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-cnzlq" Oct 04 04:51:03 crc kubenswrapper[4992]: I1004 04:51:03.470290 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-655d88ccb9-5j2nq" Oct 04 04:51:03 crc kubenswrapper[4992]: I1004 04:51:03.755324 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-579449c7d5-85jrh" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.811277 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mrgx7"] Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.816960 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.825748 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.826308 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.826348 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.826600 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wwbq8" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.833031 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mrgx7"] Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.851729 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv9nh\" (UniqueName: \"kubernetes.io/projected/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-kube-api-access-kv9nh\") pod \"dnsmasq-dns-675f4bcbfc-mrgx7\" (UID: \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.851824 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-config\") pod \"dnsmasq-dns-675f4bcbfc-mrgx7\" (UID: \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.910783 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fhhh2"] Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.912437 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.914486 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.927991 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fhhh2"] Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.953766 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-config\") pod \"dnsmasq-dns-675f4bcbfc-mrgx7\" (UID: \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.953839 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.953875 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqppr\" (UniqueName: \"kubernetes.io/projected/85f1b922-87a6-40f1-9702-3e7914c314f5-kube-api-access-gqppr\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.953950 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv9nh\" (UniqueName: \"kubernetes.io/projected/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-kube-api-access-kv9nh\") pod \"dnsmasq-dns-675f4bcbfc-mrgx7\" (UID: \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.953972 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-config\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.954615 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-config\") pod \"dnsmasq-dns-675f4bcbfc-mrgx7\" (UID: \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:20 crc kubenswrapper[4992]: I1004 04:51:20.981507 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv9nh\" (UniqueName: \"kubernetes.io/projected/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-kube-api-access-kv9nh\") pod \"dnsmasq-dns-675f4bcbfc-mrgx7\" (UID: \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.055762 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.055817 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqppr\" (UniqueName: \"kubernetes.io/projected/85f1b922-87a6-40f1-9702-3e7914c314f5-kube-api-access-gqppr\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.055875 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-config\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.056747 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-config\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.057410 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.072464 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqppr\" (UniqueName: \"kubernetes.io/projected/85f1b922-87a6-40f1-9702-3e7914c314f5-kube-api-access-gqppr\") pod \"dnsmasq-dns-78dd6ddcc-fhhh2\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.141926 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.228247 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.596253 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mrgx7"] Oct 04 04:51:21 crc kubenswrapper[4992]: W1004 04:51:21.596897 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5e6ae2a_068c_42f6_82d4_1ac949a4e06b.slice/crio-f9f1dece677ad8db548adc6212900051a35c7aabcf0ec34be97422d552b8b505 WatchSource:0}: Error finding container f9f1dece677ad8db548adc6212900051a35c7aabcf0ec34be97422d552b8b505: Status 404 returned error can't find the container with id f9f1dece677ad8db548adc6212900051a35c7aabcf0ec34be97422d552b8b505 Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.599146 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.735109 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fhhh2"] Oct 04 04:51:21 crc kubenswrapper[4992]: W1004 04:51:21.741766 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85f1b922_87a6_40f1_9702_3e7914c314f5.slice/crio-f4eaae659fe455484fc31d123b820467dac8759decd231cd41c022ea07d3bb79 WatchSource:0}: Error finding container f4eaae659fe455484fc31d123b820467dac8759decd231cd41c022ea07d3bb79: Status 404 returned error can't find the container with id f4eaae659fe455484fc31d123b820467dac8759decd231cd41c022ea07d3bb79 Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.824262 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" event={"ID":"85f1b922-87a6-40f1-9702-3e7914c314f5","Type":"ContainerStarted","Data":"f4eaae659fe455484fc31d123b820467dac8759decd231cd41c022ea07d3bb79"} Oct 04 04:51:21 crc kubenswrapper[4992]: I1004 04:51:21.825716 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" event={"ID":"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b","Type":"ContainerStarted","Data":"f9f1dece677ad8db548adc6212900051a35c7aabcf0ec34be97422d552b8b505"} Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.256715 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mrgx7"] Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.285215 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fsxfg"] Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.286757 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.293220 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fsxfg"] Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.411156 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-config\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.411219 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.411270 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lg6v\" (UniqueName: \"kubernetes.io/projected/488f98dd-8667-414f-ba52-f91d9597bf4e-kube-api-access-4lg6v\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.513611 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lg6v\" (UniqueName: \"kubernetes.io/projected/488f98dd-8667-414f-ba52-f91d9597bf4e-kube-api-access-4lg6v\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.513920 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-config\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.515475 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-config\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.515693 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.515807 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.555179 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lg6v\" (UniqueName: \"kubernetes.io/projected/488f98dd-8667-414f-ba52-f91d9597bf4e-kube-api-access-4lg6v\") pod \"dnsmasq-dns-666b6646f7-fsxfg\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.571559 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fhhh2"] Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.595481 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wkkkm"] Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.598681 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.612488 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wkkkm"] Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.620262 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.721720 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.721814 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2b6g\" (UniqueName: \"kubernetes.io/projected/a6281a12-b2ee-47e8-86e2-2550f444fd37-kube-api-access-k2b6g\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.721877 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-config\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.824568 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.824627 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2b6g\" (UniqueName: \"kubernetes.io/projected/a6281a12-b2ee-47e8-86e2-2550f444fd37-kube-api-access-k2b6g\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.824686 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-config\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.825984 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-config\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.826166 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.846035 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2b6g\" (UniqueName: \"kubernetes.io/projected/a6281a12-b2ee-47e8-86e2-2550f444fd37-kube-api-access-k2b6g\") pod \"dnsmasq-dns-57d769cc4f-wkkkm\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:24 crc kubenswrapper[4992]: I1004 04:51:24.920057 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.436536 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.437868 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.439672 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.439971 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h2k25" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.441172 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.441331 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.441556 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.441604 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.441631 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.456613 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.537309 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.537392 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.537435 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.537457 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c7fa3bc-a429-4f65-aac7-a08293b802eb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.537488 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.537511 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-config-data\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.537529 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.538639 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.538723 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzp7z\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-kube-api-access-nzp7z\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.538815 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.538846 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c7fa3bc-a429-4f65-aac7-a08293b802eb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.640777 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.640889 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.640931 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.640954 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c7fa3bc-a429-4f65-aac7-a08293b802eb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.640983 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.641002 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-config-data\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.641019 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.641075 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.641096 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzp7z\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-kube-api-access-nzp7z\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.641127 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.641151 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c7fa3bc-a429-4f65-aac7-a08293b802eb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.641270 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.641650 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.642026 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-config-data\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.642373 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.643824 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.644299 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.646076 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.649023 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.658201 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c7fa3bc-a429-4f65-aac7-a08293b802eb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.665065 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzp7z\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-kube-api-access-nzp7z\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.669435 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.671713 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c7fa3bc-a429-4f65-aac7-a08293b802eb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.715971 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.717423 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.724717 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.725691 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-558fj" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.729129 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.729201 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.729266 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.729205 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.729348 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.737640 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.771275 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849017 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849115 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrrrv\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-kube-api-access-mrrrv\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849158 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849183 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849208 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849244 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849277 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849317 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849336 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849381 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.849416 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951221 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951267 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrrrv\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-kube-api-access-mrrrv\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951291 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951322 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951349 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951418 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951463 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951493 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951509 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951528 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.951552 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.952008 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.952074 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.952650 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.952660 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.952758 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.952928 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.955045 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.955334 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.955762 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.956733 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.968899 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrrrv\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-kube-api-access-mrrrv\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:25 crc kubenswrapper[4992]: I1004 04:51:25.976513 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:26 crc kubenswrapper[4992]: I1004 04:51:26.049143 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.202029 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.204137 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.205848 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.206394 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-q77lj" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.213338 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.213452 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.213791 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.216993 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.229350 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.286732 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.286800 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-kolla-config\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.286915 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7c0512f-849a-4178-b358-989265bddf4a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.286949 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.286981 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.286997 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.287047 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-config-data-default\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.287159 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-secrets\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.287264 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxjlt\" (UniqueName: \"kubernetes.io/projected/e7c0512f-849a-4178-b358-989265bddf4a-kube-api-access-bxjlt\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395204 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395280 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-kolla-config\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395438 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7c0512f-849a-4178-b358-989265bddf4a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395466 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395528 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395549 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395574 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-config-data-default\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395601 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-secrets\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.395634 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxjlt\" (UniqueName: \"kubernetes.io/projected/e7c0512f-849a-4178-b358-989265bddf4a-kube-api-access-bxjlt\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.403381 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.404326 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-kolla-config\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.404794 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.404907 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7c0512f-849a-4178-b358-989265bddf4a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.406106 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.406719 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7c0512f-849a-4178-b358-989265bddf4a-config-data-default\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.412826 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-secrets\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.415180 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7c0512f-849a-4178-b358-989265bddf4a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.432594 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.465849 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxjlt\" (UniqueName: \"kubernetes.io/projected/e7c0512f-849a-4178-b358-989265bddf4a-kube-api-access-bxjlt\") pod \"openstack-galera-0\" (UID: \"e7c0512f-849a-4178-b358-989265bddf4a\") " pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.542724 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.609289 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.611521 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.613527 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.614230 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-c2r6g" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.614658 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.614874 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.618235 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.700617 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6mjw\" (UniqueName: \"kubernetes.io/projected/ddfacb66-d69e-4113-993a-f64918d8fc0f-kube-api-access-t6mjw\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.700848 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.700909 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.700945 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.700965 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.701016 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddfacb66-d69e-4113-993a-f64918d8fc0f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.701135 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.701166 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.701191 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.802497 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.802829 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.802859 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.802928 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6mjw\" (UniqueName: \"kubernetes.io/projected/ddfacb66-d69e-4113-993a-f64918d8fc0f-kube-api-access-t6mjw\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.803002 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.803037 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.803064 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.803085 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.803120 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddfacb66-d69e-4113-993a-f64918d8fc0f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.803693 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddfacb66-d69e-4113-993a-f64918d8fc0f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.804276 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.804514 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.804592 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.808590 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.809191 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddfacb66-d69e-4113-993a-f64918d8fc0f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.811780 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.822241 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddfacb66-d69e-4113-993a-f64918d8fc0f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.825902 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6mjw\" (UniqueName: \"kubernetes.io/projected/ddfacb66-d69e-4113-993a-f64918d8fc0f-kube-api-access-t6mjw\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.831524 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ddfacb66-d69e-4113-993a-f64918d8fc0f\") " pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:28 crc kubenswrapper[4992]: I1004 04:51:28.962509 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.143492 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.144710 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.154947 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-xpq5p" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.155136 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.155851 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.161135 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.211623 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bghst\" (UniqueName: \"kubernetes.io/projected/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-kube-api-access-bghst\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.211677 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.211707 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.211747 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-config-data\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.212052 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-kolla-config\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.314170 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-config-data\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.314243 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-kolla-config\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.314370 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bghst\" (UniqueName: \"kubernetes.io/projected/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-kube-api-access-bghst\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.314406 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.314436 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.315155 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-config-data\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.315862 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-kolla-config\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.322028 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.322211 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.333838 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bghst\" (UniqueName: \"kubernetes.io/projected/0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6-kube-api-access-bghst\") pod \"memcached-0\" (UID: \"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6\") " pod="openstack/memcached-0" Oct 04 04:51:29 crc kubenswrapper[4992]: I1004 04:51:29.473839 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 04:51:31 crc kubenswrapper[4992]: I1004 04:51:31.012711 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:51:31 crc kubenswrapper[4992]: I1004 04:51:31.014207 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:51:31 crc kubenswrapper[4992]: I1004 04:51:31.017999 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-phtgf" Oct 04 04:51:31 crc kubenswrapper[4992]: I1004 04:51:31.024297 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:51:31 crc kubenswrapper[4992]: I1004 04:51:31.140050 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncv8f\" (UniqueName: \"kubernetes.io/projected/ee9fdfcb-55b1-429c-bf7b-648b8d113c79-kube-api-access-ncv8f\") pod \"kube-state-metrics-0\" (UID: \"ee9fdfcb-55b1-429c-bf7b-648b8d113c79\") " pod="openstack/kube-state-metrics-0" Oct 04 04:51:31 crc kubenswrapper[4992]: I1004 04:51:31.242398 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncv8f\" (UniqueName: \"kubernetes.io/projected/ee9fdfcb-55b1-429c-bf7b-648b8d113c79-kube-api-access-ncv8f\") pod \"kube-state-metrics-0\" (UID: \"ee9fdfcb-55b1-429c-bf7b-648b8d113c79\") " pod="openstack/kube-state-metrics-0" Oct 04 04:51:31 crc kubenswrapper[4992]: I1004 04:51:31.266303 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncv8f\" (UniqueName: \"kubernetes.io/projected/ee9fdfcb-55b1-429c-bf7b-648b8d113c79-kube-api-access-ncv8f\") pod \"kube-state-metrics-0\" (UID: \"ee9fdfcb-55b1-429c-bf7b-648b8d113c79\") " pod="openstack/kube-state-metrics-0" Oct 04 04:51:31 crc kubenswrapper[4992]: I1004 04:51:31.330718 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.396805 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8m2tq"] Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.398170 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.400156 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.400466 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.400706 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-kwkpz" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.412582 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8m2tq"] Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.431069 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-qv6w9"] Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.432741 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.456835 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-qv6w9"] Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515297 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-run\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515402 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-lib\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515434 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68517c56-23d7-4353-ae99-eb10fee07a28-scripts\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515490 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/06640053-739f-4a61-84e0-c6ed4876a249-ovn-controller-tls-certs\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515517 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lzmp\" (UniqueName: \"kubernetes.io/projected/06640053-739f-4a61-84e0-c6ed4876a249-kube-api-access-5lzmp\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515536 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-run\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515560 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-run-ovn\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515582 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-log-ovn\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515713 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52hp7\" (UniqueName: \"kubernetes.io/projected/68517c56-23d7-4353-ae99-eb10fee07a28-kube-api-access-52hp7\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515762 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-etc-ovs\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515812 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06640053-739f-4a61-84e0-c6ed4876a249-scripts\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515848 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06640053-739f-4a61-84e0-c6ed4876a249-combined-ca-bundle\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.515917 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-log\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617301 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-log\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617417 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-run\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617451 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-lib\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617470 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68517c56-23d7-4353-ae99-eb10fee07a28-scripts\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617512 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lzmp\" (UniqueName: \"kubernetes.io/projected/06640053-739f-4a61-84e0-c6ed4876a249-kube-api-access-5lzmp\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617530 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-run\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617551 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/06640053-739f-4a61-84e0-c6ed4876a249-ovn-controller-tls-certs\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617583 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-run-ovn\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617622 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-log-ovn\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617662 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52hp7\" (UniqueName: \"kubernetes.io/projected/68517c56-23d7-4353-ae99-eb10fee07a28-kube-api-access-52hp7\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617689 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-etc-ovs\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617716 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06640053-739f-4a61-84e0-c6ed4876a249-scripts\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.617741 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06640053-739f-4a61-84e0-c6ed4876a249-combined-ca-bundle\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.618112 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-run\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.618313 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-etc-ovs\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.618352 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-log-ovn\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.618793 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-log\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.618867 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-run\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.618907 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06640053-739f-4a61-84e0-c6ed4876a249-var-run-ovn\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.620033 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68517c56-23d7-4353-ae99-eb10fee07a28-scripts\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.620260 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/68517c56-23d7-4353-ae99-eb10fee07a28-var-lib\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.620940 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06640053-739f-4a61-84e0-c6ed4876a249-scripts\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.625910 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/06640053-739f-4a61-84e0-c6ed4876a249-ovn-controller-tls-certs\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.633509 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06640053-739f-4a61-84e0-c6ed4876a249-combined-ca-bundle\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.642446 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52hp7\" (UniqueName: \"kubernetes.io/projected/68517c56-23d7-4353-ae99-eb10fee07a28-kube-api-access-52hp7\") pod \"ovn-controller-ovs-qv6w9\" (UID: \"68517c56-23d7-4353-ae99-eb10fee07a28\") " pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.646121 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lzmp\" (UniqueName: \"kubernetes.io/projected/06640053-739f-4a61-84e0-c6ed4876a249-kube-api-access-5lzmp\") pod \"ovn-controller-8m2tq\" (UID: \"06640053-739f-4a61-84e0-c6ed4876a249\") " pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.741422 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:35 crc kubenswrapper[4992]: I1004 04:51:35.754213 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.268538 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.270188 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.280260 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.280624 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.280858 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.282817 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.282993 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.283148 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-kzgcs" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.327989 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-config\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.328062 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.328090 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.328122 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qk9s\" (UniqueName: \"kubernetes.io/projected/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-kube-api-access-2qk9s\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.328145 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.328163 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.328180 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.328198 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.430813 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-config\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.430925 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.430978 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.431055 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qk9s\" (UniqueName: \"kubernetes.io/projected/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-kube-api-access-2qk9s\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.431092 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.431120 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.431146 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.431178 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.431908 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-config\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.432942 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.435322 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.435540 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.437098 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.440067 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.443473 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.457038 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qk9s\" (UniqueName: \"kubernetes.io/projected/ae844ba9-4e34-4196-9ceb-17ecec8a6ce9-kube-api-access-2qk9s\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.461174 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9\") " pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: I1004 04:51:36.603476 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:36 crc kubenswrapper[4992]: E1004 04:51:36.878292 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 04:51:36 crc kubenswrapper[4992]: E1004 04:51:36.878545 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqppr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-fhhh2_openstack(85f1b922-87a6-40f1-9702-3e7914c314f5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:51:36 crc kubenswrapper[4992]: E1004 04:51:36.879746 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" podUID="85f1b922-87a6-40f1-9702-3e7914c314f5" Oct 04 04:51:36 crc kubenswrapper[4992]: E1004 04:51:36.894198 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 04:51:36 crc kubenswrapper[4992]: E1004 04:51:36.894353 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kv9nh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-mrgx7_openstack(b5e6ae2a-068c-42f6-82d4-1ac949a4e06b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:51:36 crc kubenswrapper[4992]: E1004 04:51:36.895576 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" podUID="b5e6ae2a-068c-42f6-82d4-1ac949a4e06b" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.602023 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.665035 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-config\") pod \"85f1b922-87a6-40f1-9702-3e7914c314f5\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.665149 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-dns-svc\") pod \"85f1b922-87a6-40f1-9702-3e7914c314f5\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.665339 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqppr\" (UniqueName: \"kubernetes.io/projected/85f1b922-87a6-40f1-9702-3e7914c314f5-kube-api-access-gqppr\") pod \"85f1b922-87a6-40f1-9702-3e7914c314f5\" (UID: \"85f1b922-87a6-40f1-9702-3e7914c314f5\") " Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.667010 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-config" (OuterVolumeSpecName: "config") pod "85f1b922-87a6-40f1-9702-3e7914c314f5" (UID: "85f1b922-87a6-40f1-9702-3e7914c314f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.667446 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "85f1b922-87a6-40f1-9702-3e7914c314f5" (UID: "85f1b922-87a6-40f1-9702-3e7914c314f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.673318 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85f1b922-87a6-40f1-9702-3e7914c314f5-kube-api-access-gqppr" (OuterVolumeSpecName: "kube-api-access-gqppr") pod "85f1b922-87a6-40f1-9702-3e7914c314f5" (UID: "85f1b922-87a6-40f1-9702-3e7914c314f5"). InnerVolumeSpecName "kube-api-access-gqppr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.767764 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.768054 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqppr\" (UniqueName: \"kubernetes.io/projected/85f1b922-87a6-40f1-9702-3e7914c314f5-kube-api-access-gqppr\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.768065 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85f1b922-87a6-40f1-9702-3e7914c314f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.876683 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.908693 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.931263 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-qv6w9"] Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.940753 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.956098 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.970502 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv9nh\" (UniqueName: \"kubernetes.io/projected/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-kube-api-access-kv9nh\") pod \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\" (UID: \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\") " Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.970798 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-config\") pod \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\" (UID: \"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b\") " Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.971298 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-config" (OuterVolumeSpecName: "config") pod "b5e6ae2a-068c-42f6-82d4-1ac949a4e06b" (UID: "b5e6ae2a-068c-42f6-82d4-1ac949a4e06b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:51:37 crc kubenswrapper[4992]: I1004 04:51:37.976123 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-kube-api-access-kv9nh" (OuterVolumeSpecName: "kube-api-access-kv9nh") pod "b5e6ae2a-068c-42f6-82d4-1ac949a4e06b" (UID: "b5e6ae2a-068c-42f6-82d4-1ac949a4e06b"). InnerVolumeSpecName "kube-api-access-kv9nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.035326 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" event={"ID":"85f1b922-87a6-40f1-9702-3e7914c314f5","Type":"ContainerDied","Data":"f4eaae659fe455484fc31d123b820467dac8759decd231cd41c022ea07d3bb79"} Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.035554 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-fhhh2" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.038661 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e7c0512f-849a-4178-b358-989265bddf4a","Type":"ContainerStarted","Data":"6d7e0171ec6798ce39c05557cf6b53079da270304e28dbbbcb5d751e81489492"} Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.039445 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0c7fa3bc-a429-4f65-aac7-a08293b802eb","Type":"ContainerStarted","Data":"a8ebe42e01a8453829f66bffc8b9ce8022fe9017b94dea8cbae81582316cbcd6"} Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.040459 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qv6w9" event={"ID":"68517c56-23d7-4353-ae99-eb10fee07a28","Type":"ContainerStarted","Data":"d30eec995f86eae9fed1cd59aebc1457b4b8680f378bec5b9f58639eaa346343"} Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.044170 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" event={"ID":"b5e6ae2a-068c-42f6-82d4-1ac949a4e06b","Type":"ContainerDied","Data":"f9f1dece677ad8db548adc6212900051a35c7aabcf0ec34be97422d552b8b505"} Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.044234 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-mrgx7" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.045508 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ddfacb66-d69e-4113-993a-f64918d8fc0f","Type":"ContainerStarted","Data":"a4e2860c0d6bf6e6872e9892a9604f9f0643e9cfbb47800e5653294577ea141f"} Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.077400 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.077476 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kv9nh\" (UniqueName: \"kubernetes.io/projected/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b-kube-api-access-kv9nh\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.116532 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fhhh2"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.122836 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-fhhh2"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.137599 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mrgx7"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.142205 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-mrgx7"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.189433 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 04:51:38 crc kubenswrapper[4992]: E1004 04:51:38.212281 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85f1b922_87a6_40f1_9702_3e7914c314f5.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.219903 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.240026 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fsxfg"] Oct 04 04:51:38 crc kubenswrapper[4992]: W1004 04:51:38.259053 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6281a12_b2ee_47e8_86e2_2550f444fd37.slice/crio-de5aae0234e7a373ed13892392d58088cd13b42b8c48f3422488056521b8147f WatchSource:0}: Error finding container de5aae0234e7a373ed13892392d58088cd13b42b8c48f3422488056521b8147f: Status 404 returned error can't find the container with id de5aae0234e7a373ed13892392d58088cd13b42b8c48f3422488056521b8147f Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.294225 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8m2tq"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.309299 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wkkkm"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.345296 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85f1b922-87a6-40f1-9702-3e7914c314f5" path="/var/lib/kubelet/pods/85f1b922-87a6-40f1-9702-3e7914c314f5/volumes" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.345811 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5e6ae2a-068c-42f6-82d4-1ac949a4e06b" path="/var/lib/kubelet/pods/b5e6ae2a-068c-42f6-82d4-1ac949a4e06b/volumes" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.346240 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.346351 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-gbrkk"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.347718 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.350306 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.354800 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gbrkk"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.384467 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-config\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.384657 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-combined-ca-bundle\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.384768 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjl29\" (UniqueName: \"kubernetes.io/projected/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-kube-api-access-rjl29\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.384873 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.385008 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-ovn-rundir\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.387079 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-ovs-rundir\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.399902 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.488301 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-ovs-rundir\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.488500 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-config\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.488527 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-combined-ca-bundle\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.488558 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjl29\" (UniqueName: \"kubernetes.io/projected/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-kube-api-access-rjl29\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.488587 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.488611 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-ovn-rundir\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.488986 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-ovs-rundir\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.489738 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-config\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.489960 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-ovn-rundir\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.494346 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.498416 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-combined-ca-bundle\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.508000 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjl29\" (UniqueName: \"kubernetes.io/projected/a564ac8c-42b9-48b3-9bd3-9785ab05b19d-kube-api-access-rjl29\") pod \"ovn-controller-metrics-gbrkk\" (UID: \"a564ac8c-42b9-48b3-9bd3-9785ab05b19d\") " pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.619936 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.622217 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.628648 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.629098 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.629703 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.629862 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-7p22j" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.640973 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.676467 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gbrkk" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.691077 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.691133 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.691169 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2338a74-bd09-4ef9-b79b-6f85817977ce-config\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.691189 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.691610 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.691738 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2338a74-bd09-4ef9-b79b-6f85817977ce-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.691928 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2338a74-bd09-4ef9-b79b-6f85817977ce-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.692019 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrqsf\" (UniqueName: \"kubernetes.io/projected/f2338a74-bd09-4ef9-b79b-6f85817977ce-kube-api-access-vrqsf\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793449 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793498 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2338a74-bd09-4ef9-b79b-6f85817977ce-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793528 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2338a74-bd09-4ef9-b79b-6f85817977ce-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793590 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrqsf\" (UniqueName: \"kubernetes.io/projected/f2338a74-bd09-4ef9-b79b-6f85817977ce-kube-api-access-vrqsf\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793643 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793674 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793705 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2338a74-bd09-4ef9-b79b-6f85817977ce-config\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793763 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.793809 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.794015 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2338a74-bd09-4ef9-b79b-6f85817977ce-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.794598 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2338a74-bd09-4ef9-b79b-6f85817977ce-config\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.795576 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2338a74-bd09-4ef9-b79b-6f85817977ce-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.805182 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.805200 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.812263 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrqsf\" (UniqueName: \"kubernetes.io/projected/f2338a74-bd09-4ef9-b79b-6f85817977ce-kube-api-access-vrqsf\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.818521 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2338a74-bd09-4ef9-b79b-6f85817977ce-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.830991 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f2338a74-bd09-4ef9-b79b-6f85817977ce\") " pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:38 crc kubenswrapper[4992]: I1004 04:51:38.954540 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.062299 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ee9fdfcb-55b1-429c-bf7b-648b8d113c79","Type":"ContainerStarted","Data":"3e8024c4952b7937aa08d356ab51d3b60f23c8fbf0955f6ee5d152ae9a249f97"} Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.068801 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" event={"ID":"a6281a12-b2ee-47e8-86e2-2550f444fd37","Type":"ContainerStarted","Data":"de5aae0234e7a373ed13892392d58088cd13b42b8c48f3422488056521b8147f"} Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.071317 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9","Type":"ContainerStarted","Data":"42e84bdecb55e3b60262ae2b66616b3aea19efb4d85ba2697cc14427137f8f52"} Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.079620 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6","Type":"ContainerStarted","Data":"e952bff5172f2f117be4e33824f1cbfd58f3928df23340780f6699245a43266e"} Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.090764 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq" event={"ID":"06640053-739f-4a61-84e0-c6ed4876a249","Type":"ContainerStarted","Data":"9c7a48e1e4bc5025def0804ba89601a6e283432adb1aa3601588def9441430b8"} Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.095638 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52","Type":"ContainerStarted","Data":"2369543569a5e75a9a5f226f2abf34b54449cb743aff22d9911cb55212b8993f"} Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.099194 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" event={"ID":"488f98dd-8667-414f-ba52-f91d9597bf4e","Type":"ContainerStarted","Data":"96bb092e312dff7942e4984e9adb11d3ccdac86cca10ee2f8dc73c2fa330096c"} Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.220707 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gbrkk"] Oct 04 04:51:39 crc kubenswrapper[4992]: W1004 04:51:39.256527 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda564ac8c_42b9_48b3_9bd3_9785ab05b19d.slice/crio-801b978e18804ca990d3b79f8b40dcf1f899e264440a571fd3c2ee4660ce479c WatchSource:0}: Error finding container 801b978e18804ca990d3b79f8b40dcf1f899e264440a571fd3c2ee4660ce479c: Status 404 returned error can't find the container with id 801b978e18804ca990d3b79f8b40dcf1f899e264440a571fd3c2ee4660ce479c Oct 04 04:51:39 crc kubenswrapper[4992]: I1004 04:51:39.560521 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 04:51:40 crc kubenswrapper[4992]: I1004 04:51:40.109997 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gbrkk" event={"ID":"a564ac8c-42b9-48b3-9bd3-9785ab05b19d","Type":"ContainerStarted","Data":"801b978e18804ca990d3b79f8b40dcf1f899e264440a571fd3c2ee4660ce479c"} Oct 04 04:51:40 crc kubenswrapper[4992]: I1004 04:51:40.112859 4992 generic.go:334] "Generic (PLEG): container finished" podID="488f98dd-8667-414f-ba52-f91d9597bf4e" containerID="888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801" exitCode=0 Oct 04 04:51:40 crc kubenswrapper[4992]: I1004 04:51:40.112933 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" event={"ID":"488f98dd-8667-414f-ba52-f91d9597bf4e","Type":"ContainerDied","Data":"888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801"} Oct 04 04:51:40 crc kubenswrapper[4992]: I1004 04:51:40.134179 4992 generic.go:334] "Generic (PLEG): container finished" podID="a6281a12-b2ee-47e8-86e2-2550f444fd37" containerID="167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48" exitCode=0 Oct 04 04:51:40 crc kubenswrapper[4992]: I1004 04:51:40.134236 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" event={"ID":"a6281a12-b2ee-47e8-86e2-2550f444fd37","Type":"ContainerDied","Data":"167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48"} Oct 04 04:51:41 crc kubenswrapper[4992]: W1004 04:51:41.738041 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2338a74_bd09_4ef9_b79b_6f85817977ce.slice/crio-f6917c157f70a4edfee116e0ae195387f15cb3e9992a87d27219ae80f0996c4b WatchSource:0}: Error finding container f6917c157f70a4edfee116e0ae195387f15cb3e9992a87d27219ae80f0996c4b: Status 404 returned error can't find the container with id f6917c157f70a4edfee116e0ae195387f15cb3e9992a87d27219ae80f0996c4b Oct 04 04:51:42 crc kubenswrapper[4992]: I1004 04:51:42.146516 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f2338a74-bd09-4ef9-b79b-6f85817977ce","Type":"ContainerStarted","Data":"f6917c157f70a4edfee116e0ae195387f15cb3e9992a87d27219ae80f0996c4b"} Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.210063 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6","Type":"ContainerStarted","Data":"4b0a1112795a0511687bb1dc38e76ea9232ff9f6606aecd95f3cfd141c9734b1"} Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.210657 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.213429 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" event={"ID":"488f98dd-8667-414f-ba52-f91d9597bf4e","Type":"ContainerStarted","Data":"20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c"} Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.213842 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.215582 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f2338a74-bd09-4ef9-b79b-6f85817977ce","Type":"ContainerStarted","Data":"20596f61273997a70a05fafff33357d41cd118b1ff3c6d445b2b0b3a70f8803b"} Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.218319 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" event={"ID":"a6281a12-b2ee-47e8-86e2-2550f444fd37","Type":"ContainerStarted","Data":"ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452"} Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.218453 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.220239 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9","Type":"ContainerStarted","Data":"997fc2748ecd4574fc84c2ef9db1c9ff2b8fd34350f9c412e7d27ef713f5f27b"} Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.232764 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.496370881 podStartE2EDuration="21.23274533s" podCreationTimestamp="2025-10-04 04:51:29 +0000 UTC" firstStartedPulling="2025-10-04 04:51:38.202690296 +0000 UTC m=+1132.050365764" lastFinishedPulling="2025-10-04 04:51:46.939064755 +0000 UTC m=+1140.786740213" observedRunningTime="2025-10-04 04:51:50.22748369 +0000 UTC m=+1144.075159158" watchObservedRunningTime="2025-10-04 04:51:50.23274533 +0000 UTC m=+1144.080420798" Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.256686 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" podStartSLOduration=25.668353418 podStartE2EDuration="26.256666587s" podCreationTimestamp="2025-10-04 04:51:24 +0000 UTC" firstStartedPulling="2025-10-04 04:51:38.233614979 +0000 UTC m=+1132.081290447" lastFinishedPulling="2025-10-04 04:51:38.821928148 +0000 UTC m=+1132.669603616" observedRunningTime="2025-10-04 04:51:50.248315185 +0000 UTC m=+1144.095990673" watchObservedRunningTime="2025-10-04 04:51:50.256666587 +0000 UTC m=+1144.104342055" Oct 04 04:51:50 crc kubenswrapper[4992]: I1004 04:51:50.271530 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" podStartSLOduration=25.72517396 podStartE2EDuration="26.271488691s" podCreationTimestamp="2025-10-04 04:51:24 +0000 UTC" firstStartedPulling="2025-10-04 04:51:38.276673525 +0000 UTC m=+1132.124348993" lastFinishedPulling="2025-10-04 04:51:38.822988256 +0000 UTC m=+1132.670663724" observedRunningTime="2025-10-04 04:51:50.265093891 +0000 UTC m=+1144.112769359" watchObservedRunningTime="2025-10-04 04:51:50.271488691 +0000 UTC m=+1144.119164189" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.233586 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gbrkk" event={"ID":"a564ac8c-42b9-48b3-9bd3-9785ab05b19d","Type":"ContainerStarted","Data":"e4131d1cbb4e851963f7c6e7f2db7b8b647336c076013a61525bc7389f1b0c81"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.238026 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qv6w9" event={"ID":"68517c56-23d7-4353-ae99-eb10fee07a28","Type":"ContainerStarted","Data":"1aa7fe4413c19aac748ee37958d24df7b01ddb7b31938213f7babf1f5f7ffdb5"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.240675 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ae844ba9-4e34-4196-9ceb-17ecec8a6ce9","Type":"ContainerStarted","Data":"bc0bc5c8fc7e57bf49c368c01c369a02e49a585e1324b2ba6cf70da866185c85"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.243832 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ddfacb66-d69e-4113-993a-f64918d8fc0f","Type":"ContainerStarted","Data":"a38576f5f0322ab22076a8ad6e44b6f95bcb7dec7d843abab8a85ee54bd3fc85"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.246209 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e7c0512f-849a-4178-b358-989265bddf4a","Type":"ContainerStarted","Data":"47d29933fe055255d614ff607bb653120b6c186d8faadf5be49d76552daac4c7"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.249105 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0c7fa3bc-a429-4f65-aac7-a08293b802eb","Type":"ContainerStarted","Data":"f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.256278 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-gbrkk" podStartSLOduration=2.655119892 podStartE2EDuration="13.256260063s" podCreationTimestamp="2025-10-04 04:51:38 +0000 UTC" firstStartedPulling="2025-10-04 04:51:39.262100654 +0000 UTC m=+1133.109776132" lastFinishedPulling="2025-10-04 04:51:49.863240835 +0000 UTC m=+1143.710916303" observedRunningTime="2025-10-04 04:51:51.252233835 +0000 UTC m=+1145.099909303" watchObservedRunningTime="2025-10-04 04:51:51.256260063 +0000 UTC m=+1145.103935531" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.266875 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f2338a74-bd09-4ef9-b79b-6f85817977ce","Type":"ContainerStarted","Data":"7d3b4d610bb282ace991b812a125c1cb73b2a0bf84560d71862f3ebbb066b003"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.269435 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ee9fdfcb-55b1-429c-bf7b-648b8d113c79","Type":"ContainerStarted","Data":"cab72e8d520cea797fd7ac44a178a47c14d70f36e382b4a489f6f665cd7b2c05"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.269639 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.272337 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq" event={"ID":"06640053-739f-4a61-84e0-c6ed4876a249","Type":"ContainerStarted","Data":"7ffd41d0064664eb32b168ab8082c86c2329a2c9d648628c2513930e7c71fbaf"} Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.272405 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-8m2tq" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.347394 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.409060181 podStartE2EDuration="16.347375638s" podCreationTimestamp="2025-10-04 04:51:35 +0000 UTC" firstStartedPulling="2025-10-04 04:51:38.386246741 +0000 UTC m=+1132.233922219" lastFinishedPulling="2025-10-04 04:51:49.324562208 +0000 UTC m=+1143.172237676" observedRunningTime="2025-10-04 04:51:51.3087746 +0000 UTC m=+1145.156450078" watchObservedRunningTime="2025-10-04 04:51:51.347375638 +0000 UTC m=+1145.195051106" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.431375 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=6.850974203 podStartE2EDuration="14.431347753s" podCreationTimestamp="2025-10-04 04:51:37 +0000 UTC" firstStartedPulling="2025-10-04 04:51:41.744524076 +0000 UTC m=+1135.592199554" lastFinishedPulling="2025-10-04 04:51:49.324897626 +0000 UTC m=+1143.172573104" observedRunningTime="2025-10-04 04:51:51.427998174 +0000 UTC m=+1145.275673642" watchObservedRunningTime="2025-10-04 04:51:51.431347753 +0000 UTC m=+1145.279023221" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.450078 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.773914906 podStartE2EDuration="21.450062081s" podCreationTimestamp="2025-10-04 04:51:30 +0000 UTC" firstStartedPulling="2025-10-04 04:51:38.243954884 +0000 UTC m=+1132.091630352" lastFinishedPulling="2025-10-04 04:51:49.920102059 +0000 UTC m=+1143.767777527" observedRunningTime="2025-10-04 04:51:51.444304868 +0000 UTC m=+1145.291980336" watchObservedRunningTime="2025-10-04 04:51:51.450062081 +0000 UTC m=+1145.297737549" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.463958 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8m2tq" podStartSLOduration=5.410005546 podStartE2EDuration="16.46394115s" podCreationTimestamp="2025-10-04 04:51:35 +0000 UTC" firstStartedPulling="2025-10-04 04:51:38.270981633 +0000 UTC m=+1132.118657101" lastFinishedPulling="2025-10-04 04:51:49.324917237 +0000 UTC m=+1143.172592705" observedRunningTime="2025-10-04 04:51:51.462488222 +0000 UTC m=+1145.310163700" watchObservedRunningTime="2025-10-04 04:51:51.46394115 +0000 UTC m=+1145.311616618" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.508558 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wkkkm"] Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.545952 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8k5hc"] Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.547877 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.552714 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.561090 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8k5hc"] Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.604103 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.604193 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.655185 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.655237 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.655300 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-config\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.655432 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9txjb\" (UniqueName: \"kubernetes.io/projected/9d90c8e7-c146-4390-a587-3bd054f59831-kube-api-access-9txjb\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.756797 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.757042 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.757170 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-config\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.757384 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9txjb\" (UniqueName: \"kubernetes.io/projected/9d90c8e7-c146-4390-a587-3bd054f59831-kube-api-access-9txjb\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.757832 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.758529 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-config\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.758757 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.800496 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9txjb\" (UniqueName: \"kubernetes.io/projected/9d90c8e7-c146-4390-a587-3bd054f59831-kube-api-access-9txjb\") pod \"dnsmasq-dns-7fd796d7df-8k5hc\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.826644 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fsxfg"] Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.869548 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.918449 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rdf2t"] Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.919892 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.923094 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.941343 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rdf2t"] Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.963428 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.963688 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-config\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.963890 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.964009 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnw5l\" (UniqueName: \"kubernetes.io/projected/ec6764f1-efcc-4dac-a681-0af9276accdd-kube-api-access-nnw5l\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:51 crc kubenswrapper[4992]: I1004 04:51:51.964124 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.065984 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.066067 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-config\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.066101 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.066149 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnw5l\" (UniqueName: \"kubernetes.io/projected/ec6764f1-efcc-4dac-a681-0af9276accdd-kube-api-access-nnw5l\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.066218 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.067245 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.067898 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.068549 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-config\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.069336 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.097275 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnw5l\" (UniqueName: \"kubernetes.io/projected/ec6764f1-efcc-4dac-a681-0af9276accdd-kube-api-access-nnw5l\") pod \"dnsmasq-dns-86db49b7ff-rdf2t\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.280107 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52","Type":"ContainerStarted","Data":"c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44"} Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.281716 4992 generic.go:334] "Generic (PLEG): container finished" podID="68517c56-23d7-4353-ae99-eb10fee07a28" containerID="1aa7fe4413c19aac748ee37958d24df7b01ddb7b31938213f7babf1f5f7ffdb5" exitCode=0 Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.281799 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qv6w9" event={"ID":"68517c56-23d7-4353-ae99-eb10fee07a28","Type":"ContainerDied","Data":"1aa7fe4413c19aac748ee37958d24df7b01ddb7b31938213f7babf1f5f7ffdb5"} Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.282176 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" podUID="488f98dd-8667-414f-ba52-f91d9597bf4e" containerName="dnsmasq-dns" containerID="cri-o://20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c" gracePeriod=10 Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.283869 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" podUID="a6281a12-b2ee-47e8-86e2-2550f444fd37" containerName="dnsmasq-dns" containerID="cri-o://ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452" gracePeriod=10 Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.284200 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.362294 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8k5hc"] Oct 04 04:51:52 crc kubenswrapper[4992]: W1004 04:51:52.450341 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d90c8e7_c146_4390_a587_3bd054f59831.slice/crio-148244385d56e5865e8289895da10e97886f4e12d4d06a1bbad07c7222aaac46 WatchSource:0}: Error finding container 148244385d56e5865e8289895da10e97886f4e12d4d06a1bbad07c7222aaac46: Status 404 returned error can't find the container with id 148244385d56e5865e8289895da10e97886f4e12d4d06a1bbad07c7222aaac46 Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.590172 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rdf2t"] Oct 04 04:51:52 crc kubenswrapper[4992]: W1004 04:51:52.599851 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec6764f1_efcc_4dac_a681_0af9276accdd.slice/crio-9ae3a40fe48c645a647b423a01d93dc9434d351b884e981303ae9a1b7d6cb461 WatchSource:0}: Error finding container 9ae3a40fe48c645a647b423a01d93dc9434d351b884e981303ae9a1b7d6cb461: Status 404 returned error can't find the container with id 9ae3a40fe48c645a647b423a01d93dc9434d351b884e981303ae9a1b7d6cb461 Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.953089 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.979685 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-config\") pod \"488f98dd-8667-414f-ba52-f91d9597bf4e\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.979731 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lg6v\" (UniqueName: \"kubernetes.io/projected/488f98dd-8667-414f-ba52-f91d9597bf4e-kube-api-access-4lg6v\") pod \"488f98dd-8667-414f-ba52-f91d9597bf4e\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.979818 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-dns-svc\") pod \"488f98dd-8667-414f-ba52-f91d9597bf4e\" (UID: \"488f98dd-8667-414f-ba52-f91d9597bf4e\") " Oct 04 04:51:52 crc kubenswrapper[4992]: I1004 04:51:52.989024 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/488f98dd-8667-414f-ba52-f91d9597bf4e-kube-api-access-4lg6v" (OuterVolumeSpecName: "kube-api-access-4lg6v") pod "488f98dd-8667-414f-ba52-f91d9597bf4e" (UID: "488f98dd-8667-414f-ba52-f91d9597bf4e"). InnerVolumeSpecName "kube-api-access-4lg6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.049403 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-config" (OuterVolumeSpecName: "config") pod "488f98dd-8667-414f-ba52-f91d9597bf4e" (UID: "488f98dd-8667-414f-ba52-f91d9597bf4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.053708 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.080513 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2b6g\" (UniqueName: \"kubernetes.io/projected/a6281a12-b2ee-47e8-86e2-2550f444fd37-kube-api-access-k2b6g\") pod \"a6281a12-b2ee-47e8-86e2-2550f444fd37\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.080597 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-config\") pod \"a6281a12-b2ee-47e8-86e2-2550f444fd37\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.080630 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-dns-svc\") pod \"a6281a12-b2ee-47e8-86e2-2550f444fd37\" (UID: \"a6281a12-b2ee-47e8-86e2-2550f444fd37\") " Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.080837 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.080848 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lg6v\" (UniqueName: \"kubernetes.io/projected/488f98dd-8667-414f-ba52-f91d9597bf4e-kube-api-access-4lg6v\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.088016 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6281a12-b2ee-47e8-86e2-2550f444fd37-kube-api-access-k2b6g" (OuterVolumeSpecName: "kube-api-access-k2b6g") pod "a6281a12-b2ee-47e8-86e2-2550f444fd37" (UID: "a6281a12-b2ee-47e8-86e2-2550f444fd37"). InnerVolumeSpecName "kube-api-access-k2b6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.105752 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "488f98dd-8667-414f-ba52-f91d9597bf4e" (UID: "488f98dd-8667-414f-ba52-f91d9597bf4e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.124396 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6281a12-b2ee-47e8-86e2-2550f444fd37" (UID: "a6281a12-b2ee-47e8-86e2-2550f444fd37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.142777 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-config" (OuterVolumeSpecName: "config") pod "a6281a12-b2ee-47e8-86e2-2550f444fd37" (UID: "a6281a12-b2ee-47e8-86e2-2550f444fd37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.181823 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2b6g\" (UniqueName: \"kubernetes.io/projected/a6281a12-b2ee-47e8-86e2-2550f444fd37-kube-api-access-k2b6g\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.181858 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/488f98dd-8667-414f-ba52-f91d9597bf4e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.181867 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.181875 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6281a12-b2ee-47e8-86e2-2550f444fd37-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.289169 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" event={"ID":"ec6764f1-efcc-4dac-a681-0af9276accdd","Type":"ContainerStarted","Data":"9ae3a40fe48c645a647b423a01d93dc9434d351b884e981303ae9a1b7d6cb461"} Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.291661 4992 generic.go:334] "Generic (PLEG): container finished" podID="488f98dd-8667-414f-ba52-f91d9597bf4e" containerID="20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c" exitCode=0 Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.291722 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" event={"ID":"488f98dd-8667-414f-ba52-f91d9597bf4e","Type":"ContainerDied","Data":"20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c"} Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.291751 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" event={"ID":"488f98dd-8667-414f-ba52-f91d9597bf4e","Type":"ContainerDied","Data":"96bb092e312dff7942e4984e9adb11d3ccdac86cca10ee2f8dc73c2fa330096c"} Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.291773 4992 scope.go:117] "RemoveContainer" containerID="20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.291883 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fsxfg" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.300194 4992 generic.go:334] "Generic (PLEG): container finished" podID="9d90c8e7-c146-4390-a587-3bd054f59831" containerID="013ad68102819b8f3c7334c57b21a3fcaea70082cdacc491728bdaed1ccaed0f" exitCode=0 Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.300905 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" event={"ID":"9d90c8e7-c146-4390-a587-3bd054f59831","Type":"ContainerDied","Data":"013ad68102819b8f3c7334c57b21a3fcaea70082cdacc491728bdaed1ccaed0f"} Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.300935 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" event={"ID":"9d90c8e7-c146-4390-a587-3bd054f59831","Type":"ContainerStarted","Data":"148244385d56e5865e8289895da10e97886f4e12d4d06a1bbad07c7222aaac46"} Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.306921 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qv6w9" event={"ID":"68517c56-23d7-4353-ae99-eb10fee07a28","Type":"ContainerStarted","Data":"30e6e9b104e89be67fa6af75a561b6d3a2d6c9c6498a7a17510df8c0ba7beb52"} Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.312646 4992 generic.go:334] "Generic (PLEG): container finished" podID="a6281a12-b2ee-47e8-86e2-2550f444fd37" containerID="ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452" exitCode=0 Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.313778 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.315257 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" event={"ID":"a6281a12-b2ee-47e8-86e2-2550f444fd37","Type":"ContainerDied","Data":"ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452"} Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.315291 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-wkkkm" event={"ID":"a6281a12-b2ee-47e8-86e2-2550f444fd37","Type":"ContainerDied","Data":"de5aae0234e7a373ed13892392d58088cd13b42b8c48f3422488056521b8147f"} Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.329577 4992 scope.go:117] "RemoveContainer" containerID="888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.352196 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fsxfg"] Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.361712 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fsxfg"] Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.367695 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wkkkm"] Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.372799 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-wkkkm"] Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.956288 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.956585 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:53 crc kubenswrapper[4992]: I1004 04:51:53.997764 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:54 crc kubenswrapper[4992]: I1004 04:51:54.335958 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="488f98dd-8667-414f-ba52-f91d9597bf4e" path="/var/lib/kubelet/pods/488f98dd-8667-414f-ba52-f91d9597bf4e/volumes" Oct 04 04:51:54 crc kubenswrapper[4992]: I1004 04:51:54.336684 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6281a12-b2ee-47e8-86e2-2550f444fd37" path="/var/lib/kubelet/pods/a6281a12-b2ee-47e8-86e2-2550f444fd37/volumes" Oct 04 04:51:54 crc kubenswrapper[4992]: I1004 04:51:54.643280 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:55 crc kubenswrapper[4992]: I1004 04:51:55.396082 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.646252 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.776234 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:51:56 crc kubenswrapper[4992]: E1004 04:51:56.776627 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="488f98dd-8667-414f-ba52-f91d9597bf4e" containerName="dnsmasq-dns" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.776641 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="488f98dd-8667-414f-ba52-f91d9597bf4e" containerName="dnsmasq-dns" Oct 04 04:51:56 crc kubenswrapper[4992]: E1004 04:51:56.776660 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6281a12-b2ee-47e8-86e2-2550f444fd37" containerName="dnsmasq-dns" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.776668 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6281a12-b2ee-47e8-86e2-2550f444fd37" containerName="dnsmasq-dns" Oct 04 04:51:56 crc kubenswrapper[4992]: E1004 04:51:56.776685 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="488f98dd-8667-414f-ba52-f91d9597bf4e" containerName="init" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.776694 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="488f98dd-8667-414f-ba52-f91d9597bf4e" containerName="init" Oct 04 04:51:56 crc kubenswrapper[4992]: E1004 04:51:56.776713 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6281a12-b2ee-47e8-86e2-2550f444fd37" containerName="init" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.776720 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6281a12-b2ee-47e8-86e2-2550f444fd37" containerName="init" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.777038 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6281a12-b2ee-47e8-86e2-2550f444fd37" containerName="dnsmasq-dns" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.777059 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="488f98dd-8667-414f-ba52-f91d9597bf4e" containerName="dnsmasq-dns" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.778093 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.782044 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.782697 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.783688 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.783802 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-fm9gt" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.795774 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.843874 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.843970 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-config\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.844002 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-scripts\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.844023 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfxss\" (UniqueName: \"kubernetes.io/projected/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-kube-api-access-pfxss\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.844044 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.844138 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.844157 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.946128 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-scripts\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.946190 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfxss\" (UniqueName: \"kubernetes.io/projected/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-kube-api-access-pfxss\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.946223 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.946288 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.946316 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.946385 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.946442 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-config\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.946875 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.947093 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-scripts\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.947347 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-config\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.952682 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.952836 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.960338 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:56 crc kubenswrapper[4992]: I1004 04:51:56.968174 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfxss\" (UniqueName: \"kubernetes.io/projected/9a31b3de-d7c5-4e1e-8ea8-fce11d130790-kube-api-access-pfxss\") pod \"ovn-northd-0\" (UID: \"9a31b3de-d7c5-4e1e-8ea8-fce11d130790\") " pod="openstack/ovn-northd-0" Oct 04 04:51:57 crc kubenswrapper[4992]: I1004 04:51:57.097976 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 04:51:58 crc kubenswrapper[4992]: I1004 04:51:58.364300 4992 generic.go:334] "Generic (PLEG): container finished" podID="ec6764f1-efcc-4dac-a681-0af9276accdd" containerID="c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8" exitCode=0 Oct 04 04:51:58 crc kubenswrapper[4992]: I1004 04:51:58.364349 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" event={"ID":"ec6764f1-efcc-4dac-a681-0af9276accdd","Type":"ContainerDied","Data":"c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8"} Oct 04 04:51:58 crc kubenswrapper[4992]: E1004 04:51:58.641561 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddfacb66_d69e_4113_993a_f64918d8fc0f.slice/crio-conmon-a38576f5f0322ab22076a8ad6e44b6f95bcb7dec7d843abab8a85ee54bd3fc85.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:51:59 crc kubenswrapper[4992]: I1004 04:51:59.376252 4992 generic.go:334] "Generic (PLEG): container finished" podID="e7c0512f-849a-4178-b358-989265bddf4a" containerID="47d29933fe055255d614ff607bb653120b6c186d8faadf5be49d76552daac4c7" exitCode=0 Oct 04 04:51:59 crc kubenswrapper[4992]: I1004 04:51:59.376336 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e7c0512f-849a-4178-b358-989265bddf4a","Type":"ContainerDied","Data":"47d29933fe055255d614ff607bb653120b6c186d8faadf5be49d76552daac4c7"} Oct 04 04:51:59 crc kubenswrapper[4992]: I1004 04:51:59.378466 4992 generic.go:334] "Generic (PLEG): container finished" podID="ddfacb66-d69e-4113-993a-f64918d8fc0f" containerID="a38576f5f0322ab22076a8ad6e44b6f95bcb7dec7d843abab8a85ee54bd3fc85" exitCode=0 Oct 04 04:51:59 crc kubenswrapper[4992]: I1004 04:51:59.378506 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ddfacb66-d69e-4113-993a-f64918d8fc0f","Type":"ContainerDied","Data":"a38576f5f0322ab22076a8ad6e44b6f95bcb7dec7d843abab8a85ee54bd3fc85"} Oct 04 04:51:59 crc kubenswrapper[4992]: I1004 04:51:59.476323 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 04 04:52:00 crc kubenswrapper[4992]: I1004 04:52:00.389524 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qv6w9" event={"ID":"68517c56-23d7-4353-ae99-eb10fee07a28","Type":"ContainerStarted","Data":"c5b107c7279583a9fd59f8e641406b4ca5328190371a7d24a40994bd177d4b96"} Oct 04 04:52:01 crc kubenswrapper[4992]: I1004 04:52:01.338072 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 04:52:01 crc kubenswrapper[4992]: I1004 04:52:01.407453 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8k5hc"] Oct 04 04:52:01 crc kubenswrapper[4992]: I1004 04:52:01.445510 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-4jk5d"] Oct 04 04:52:01 crc kubenswrapper[4992]: I1004 04:52:01.446900 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.471261 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4jk5d"] Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.553181 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.553517 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.553554 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-config\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.553586 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf9tz\" (UniqueName: \"kubernetes.io/projected/365fab3e-1695-4e8e-a362-485daebb7f46-kube-api-access-vf9tz\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.553683 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-dns-svc\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.655007 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.655072 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.655115 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-config\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.655159 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf9tz\" (UniqueName: \"kubernetes.io/projected/365fab3e-1695-4e8e-a362-485daebb7f46-kube-api-access-vf9tz\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.655206 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-dns-svc\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.656081 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.656081 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-config\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.656126 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-dns-svc\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.656124 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.678584 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf9tz\" (UniqueName: \"kubernetes.io/projected/365fab3e-1695-4e8e-a362-485daebb7f46-kube-api-access-vf9tz\") pod \"dnsmasq-dns-698758b865-4jk5d\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:01.765222 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.404686 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.404734 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.436243 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-qv6w9" podStartSLOduration=16.430624211 podStartE2EDuration="27.436225119s" podCreationTimestamp="2025-10-04 04:51:35 +0000 UTC" firstStartedPulling="2025-10-04 04:51:37.920152635 +0000 UTC m=+1131.767828103" lastFinishedPulling="2025-10-04 04:51:48.925753523 +0000 UTC m=+1142.773429011" observedRunningTime="2025-10-04 04:52:02.430786415 +0000 UTC m=+1156.278461903" watchObservedRunningTime="2025-10-04 04:52:02.436225119 +0000 UTC m=+1156.283900597" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.548070 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.555538 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.558481 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.558484 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.558709 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-m69gw" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.560510 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.623142 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.672652 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpsl5\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-kube-api-access-dpsl5\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.674661 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.674717 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/fa875baa-782c-469c-816f-7727c4885d28-cache\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.674775 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/fa875baa-782c-469c-816f-7727c4885d28-lock\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.674861 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.776815 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.777114 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/fa875baa-782c-469c-816f-7727c4885d28-cache\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.777142 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/fa875baa-782c-469c-816f-7727c4885d28-lock\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: E1004 04:52:02.777039 4992 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:52:02 crc kubenswrapper[4992]: E1004 04:52:02.777204 4992 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.777179 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: E1004 04:52:02.777265 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift podName:fa875baa-782c-469c-816f-7727c4885d28 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:03.277244076 +0000 UTC m=+1157.124919544 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift") pod "swift-storage-0" (UID: "fa875baa-782c-469c-816f-7727c4885d28") : configmap "swift-ring-files" not found Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.777327 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpsl5\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-kube-api-access-dpsl5\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.777429 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.777681 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/fa875baa-782c-469c-816f-7727c4885d28-lock\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.777813 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/fa875baa-782c-469c-816f-7727c4885d28-cache\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.798842 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:02 crc kubenswrapper[4992]: I1004 04:52:02.800087 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpsl5\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-kube-api-access-dpsl5\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:03 crc kubenswrapper[4992]: I1004 04:52:03.284788 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:03 crc kubenswrapper[4992]: E1004 04:52:03.285107 4992 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:52:03 crc kubenswrapper[4992]: E1004 04:52:03.285144 4992 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:52:03 crc kubenswrapper[4992]: E1004 04:52:03.285203 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift podName:fa875baa-782c-469c-816f-7727c4885d28 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:04.285183335 +0000 UTC m=+1158.132858803 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift") pod "swift-storage-0" (UID: "fa875baa-782c-469c-816f-7727c4885d28") : configmap "swift-ring-files" not found Oct 04 04:52:04 crc kubenswrapper[4992]: I1004 04:52:04.301950 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:04 crc kubenswrapper[4992]: E1004 04:52:04.302151 4992 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:52:04 crc kubenswrapper[4992]: E1004 04:52:04.302171 4992 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:52:04 crc kubenswrapper[4992]: E1004 04:52:04.302232 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift podName:fa875baa-782c-469c-816f-7727c4885d28 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:06.302213195 +0000 UTC m=+1160.149888683 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift") pod "swift-storage-0" (UID: "fa875baa-782c-469c-816f-7727c4885d28") : configmap "swift-ring-files" not found Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.129471 4992 scope.go:117] "RemoveContainer" containerID="20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c" Oct 04 04:52:05 crc kubenswrapper[4992]: E1004 04:52:05.130231 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c\": container with ID starting with 20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c not found: ID does not exist" containerID="20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.130301 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c"} err="failed to get container status \"20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c\": rpc error: code = NotFound desc = could not find container \"20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c\": container with ID starting with 20834ee92cb3fd0be873e91717670a767dd59dff9ef10a83c75db397afd5e69c not found: ID does not exist" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.130328 4992 scope.go:117] "RemoveContainer" containerID="888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801" Oct 04 04:52:05 crc kubenswrapper[4992]: E1004 04:52:05.130736 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801\": container with ID starting with 888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801 not found: ID does not exist" containerID="888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.130761 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801"} err="failed to get container status \"888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801\": rpc error: code = NotFound desc = could not find container \"888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801\": container with ID starting with 888188c56e2202c68b2ae8f5305de2a1896a8058fd3a579c39fb32e467d18801 not found: ID does not exist" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.130774 4992 scope.go:117] "RemoveContainer" containerID="ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.268616 4992 scope.go:117] "RemoveContainer" containerID="167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.319566 4992 scope.go:117] "RemoveContainer" containerID="ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452" Oct 04 04:52:05 crc kubenswrapper[4992]: E1004 04:52:05.321430 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452\": container with ID starting with ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452 not found: ID does not exist" containerID="ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.321576 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452"} err="failed to get container status \"ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452\": rpc error: code = NotFound desc = could not find container \"ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452\": container with ID starting with ab75d37294acc57fc096216c8f6f368f8701ce1f7549ae55d0c249d3741ff452 not found: ID does not exist" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.321618 4992 scope.go:117] "RemoveContainer" containerID="167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48" Oct 04 04:52:05 crc kubenswrapper[4992]: E1004 04:52:05.330575 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48\": container with ID starting with 167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48 not found: ID does not exist" containerID="167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.330634 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48"} err="failed to get container status \"167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48\": rpc error: code = NotFound desc = could not find container \"167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48\": container with ID starting with 167bbfe21014a8e4aeb2d627a63347af96538fded11be8b77e5714e30fa7aa48 not found: ID does not exist" Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.645265 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4jk5d"] Oct 04 04:52:05 crc kubenswrapper[4992]: W1004 04:52:05.647320 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod365fab3e_1695_4e8e_a362_485daebb7f46.slice/crio-d3174bbb8fcfefcba7d9dffaacebca5ad1c03ffad9bda4a7d2f982bc9f4e169e WatchSource:0}: Error finding container d3174bbb8fcfefcba7d9dffaacebca5ad1c03ffad9bda4a7d2f982bc9f4e169e: Status 404 returned error can't find the container with id d3174bbb8fcfefcba7d9dffaacebca5ad1c03ffad9bda4a7d2f982bc9f4e169e Oct 04 04:52:05 crc kubenswrapper[4992]: I1004 04:52:05.743253 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.343320 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:06 crc kubenswrapper[4992]: E1004 04:52:06.343568 4992 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:52:06 crc kubenswrapper[4992]: E1004 04:52:06.343601 4992 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:52:06 crc kubenswrapper[4992]: E1004 04:52:06.343685 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift podName:fa875baa-782c-469c-816f-7727c4885d28 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:10.34366165 +0000 UTC m=+1164.191337118 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift") pod "swift-storage-0" (UID: "fa875baa-782c-469c-816f-7727c4885d28") : configmap "swift-ring-files" not found Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.443443 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" event={"ID":"9d90c8e7-c146-4390-a587-3bd054f59831","Type":"ContainerStarted","Data":"ec9f8879072b8f14461e3c0430cfefcca2b8986ecb19db42e432adc35960d596"} Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.445333 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4jk5d" event={"ID":"365fab3e-1695-4e8e-a362-485daebb7f46","Type":"ContainerStarted","Data":"17e420e4d0d0c301a5c9c64116705a4065e33c585ae24d9d431da21b1784953e"} Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.445408 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4jk5d" event={"ID":"365fab3e-1695-4e8e-a362-485daebb7f46","Type":"ContainerStarted","Data":"d3174bbb8fcfefcba7d9dffaacebca5ad1c03ffad9bda4a7d2f982bc9f4e169e"} Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.447243 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ddfacb66-d69e-4113-993a-f64918d8fc0f","Type":"ContainerStarted","Data":"27693f4e6e9bdf0eef7e53b92d226f920d379838e38467d39db7f8cff0fb343c"} Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.449422 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e7c0512f-849a-4178-b358-989265bddf4a","Type":"ContainerStarted","Data":"a69f84a20158fae992fa6db114ba21b8106e707d858fd60bc680f8954430c803"} Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.451452 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" event={"ID":"ec6764f1-efcc-4dac-a681-0af9276accdd","Type":"ContainerStarted","Data":"5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef"} Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.452972 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9a31b3de-d7c5-4e1e-8ea8-fce11d130790","Type":"ContainerStarted","Data":"4d6c0aaf7073c248af3e852bc823d70e85d4a35c55f990ba256aab5b73105f7e"} Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.533151 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wvk6d"] Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.534140 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.536683 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.537567 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.538232 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.565229 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wvk6d"] Oct 04 04:52:06 crc kubenswrapper[4992]: E1004 04:52:06.566083 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-2mvnx ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-2mvnx ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-wvk6d" podUID="6bdbdac2-5df9-46ab-9965-c5970241fe90" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.572430 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9hx4r"] Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.574122 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.583962 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9hx4r"] Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.616876 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wvk6d"] Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.648597 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-dispersionconf\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.648675 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-combined-ca-bundle\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.648711 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-ring-data-devices\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.648748 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-scripts\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.648771 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mvnx\" (UniqueName: \"kubernetes.io/projected/6bdbdac2-5df9-46ab-9965-c5970241fe90-kube-api-access-2mvnx\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.648933 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-swiftconf\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.648977 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-scripts\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.649058 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-ring-data-devices\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.649162 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-dispersionconf\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.649265 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c9af92df-6216-469d-84e9-a09e3b8a648d-etc-swift\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.649410 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-combined-ca-bundle\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.649451 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h4zk\" (UniqueName: \"kubernetes.io/projected/c9af92df-6216-469d-84e9-a09e3b8a648d-kube-api-access-2h4zk\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.649492 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-swiftconf\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.649759 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6bdbdac2-5df9-46ab-9965-c5970241fe90-etc-swift\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.750694 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6bdbdac2-5df9-46ab-9965-c5970241fe90-etc-swift\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.750745 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-dispersionconf\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.750771 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-combined-ca-bundle\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.750832 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-ring-data-devices\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.751190 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6bdbdac2-5df9-46ab-9965-c5970241fe90-etc-swift\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.751809 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-scripts\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.751997 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mvnx\" (UniqueName: \"kubernetes.io/projected/6bdbdac2-5df9-46ab-9965-c5970241fe90-kube-api-access-2mvnx\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.752185 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-swiftconf\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.752239 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-scripts\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.752337 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-ring-data-devices\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.752474 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-dispersionconf\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.752580 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c9af92df-6216-469d-84e9-a09e3b8a648d-etc-swift\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.752168 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-ring-data-devices\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.753061 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-scripts\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.753551 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-scripts\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.753610 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-combined-ca-bundle\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.753658 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h4zk\" (UniqueName: \"kubernetes.io/projected/c9af92df-6216-469d-84e9-a09e3b8a648d-kube-api-access-2h4zk\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.753716 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-swiftconf\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.753762 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c9af92df-6216-469d-84e9-a09e3b8a648d-etc-swift\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.754128 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-ring-data-devices\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.756840 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-combined-ca-bundle\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.757719 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-combined-ca-bundle\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.758261 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-swiftconf\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.758917 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-swiftconf\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.760561 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-dispersionconf\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.765863 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-dispersionconf\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.775300 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mvnx\" (UniqueName: \"kubernetes.io/projected/6bdbdac2-5df9-46ab-9965-c5970241fe90-kube-api-access-2mvnx\") pod \"swift-ring-rebalance-wvk6d\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.776033 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h4zk\" (UniqueName: \"kubernetes.io/projected/c9af92df-6216-469d-84e9-a09e3b8a648d-kube-api-access-2h4zk\") pod \"swift-ring-rebalance-9hx4r\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:06 crc kubenswrapper[4992]: I1004 04:52:06.896953 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.340222 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9hx4r"] Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.463590 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9hx4r" event={"ID":"c9af92df-6216-469d-84e9-a09e3b8a648d","Type":"ContainerStarted","Data":"7c703c998c8be3fd65f5f59ebd45228e169257833266463910de7b176d7feb2a"} Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.474281 4992 generic.go:334] "Generic (PLEG): container finished" podID="365fab3e-1695-4e8e-a362-485daebb7f46" containerID="17e420e4d0d0c301a5c9c64116705a4065e33c585ae24d9d431da21b1784953e" exitCode=0 Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.474397 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4jk5d" event={"ID":"365fab3e-1695-4e8e-a362-485daebb7f46","Type":"ContainerDied","Data":"17e420e4d0d0c301a5c9c64116705a4065e33c585ae24d9d431da21b1784953e"} Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.475052 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.475139 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.475326 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" podUID="9d90c8e7-c146-4390-a587-3bd054f59831" containerName="dnsmasq-dns" containerID="cri-o://ec9f8879072b8f14461e3c0430cfefcca2b8986ecb19db42e432adc35960d596" gracePeriod=10 Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.534153 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" podStartSLOduration=16.534131756 podStartE2EDuration="16.534131756s" podCreationTimestamp="2025-10-04 04:51:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:52:07.531029454 +0000 UTC m=+1161.378704952" watchObservedRunningTime="2025-10-04 04:52:07.534131756 +0000 UTC m=+1161.381807234" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.557666 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=31.538141167 podStartE2EDuration="40.557639292s" podCreationTimestamp="2025-10-04 04:51:27 +0000 UTC" firstStartedPulling="2025-10-04 04:51:37.919785666 +0000 UTC m=+1131.767461134" lastFinishedPulling="2025-10-04 04:51:46.939283791 +0000 UTC m=+1140.786959259" observedRunningTime="2025-10-04 04:52:07.553969104 +0000 UTC m=+1161.401644602" watchObservedRunningTime="2025-10-04 04:52:07.557639292 +0000 UTC m=+1161.405314760" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.584496 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.587582 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=29.877071415 podStartE2EDuration="40.587566618s" podCreationTimestamp="2025-10-04 04:51:27 +0000 UTC" firstStartedPulling="2025-10-04 04:51:37.919550499 +0000 UTC m=+1131.767225967" lastFinishedPulling="2025-10-04 04:51:48.630045702 +0000 UTC m=+1142.477721170" observedRunningTime="2025-10-04 04:52:07.579313159 +0000 UTC m=+1161.426988657" watchObservedRunningTime="2025-10-04 04:52:07.587566618 +0000 UTC m=+1161.435242086" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.600827 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" podStartSLOduration=16.600808331 podStartE2EDuration="16.600808331s" podCreationTimestamp="2025-10-04 04:51:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:52:07.598369726 +0000 UTC m=+1161.446045194" watchObservedRunningTime="2025-10-04 04:52:07.600808331 +0000 UTC m=+1161.448483809" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.682443 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mvnx\" (UniqueName: \"kubernetes.io/projected/6bdbdac2-5df9-46ab-9965-c5970241fe90-kube-api-access-2mvnx\") pod \"6bdbdac2-5df9-46ab-9965-c5970241fe90\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.682785 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-swiftconf\") pod \"6bdbdac2-5df9-46ab-9965-c5970241fe90\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.682890 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-combined-ca-bundle\") pod \"6bdbdac2-5df9-46ab-9965-c5970241fe90\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.682983 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-ring-data-devices\") pod \"6bdbdac2-5df9-46ab-9965-c5970241fe90\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.683102 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-dispersionconf\") pod \"6bdbdac2-5df9-46ab-9965-c5970241fe90\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.683281 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-scripts\") pod \"6bdbdac2-5df9-46ab-9965-c5970241fe90\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.683466 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6bdbdac2-5df9-46ab-9965-c5970241fe90-etc-swift\") pod \"6bdbdac2-5df9-46ab-9965-c5970241fe90\" (UID: \"6bdbdac2-5df9-46ab-9965-c5970241fe90\") " Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.684463 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6bdbdac2-5df9-46ab-9965-c5970241fe90" (UID: "6bdbdac2-5df9-46ab-9965-c5970241fe90"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.684755 4992 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.684917 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-scripts" (OuterVolumeSpecName: "scripts") pod "6bdbdac2-5df9-46ab-9965-c5970241fe90" (UID: "6bdbdac2-5df9-46ab-9965-c5970241fe90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.685140 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bdbdac2-5df9-46ab-9965-c5970241fe90-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6bdbdac2-5df9-46ab-9965-c5970241fe90" (UID: "6bdbdac2-5df9-46ab-9965-c5970241fe90"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.687731 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bdbdac2-5df9-46ab-9965-c5970241fe90-kube-api-access-2mvnx" (OuterVolumeSpecName: "kube-api-access-2mvnx") pod "6bdbdac2-5df9-46ab-9965-c5970241fe90" (UID: "6bdbdac2-5df9-46ab-9965-c5970241fe90"). InnerVolumeSpecName "kube-api-access-2mvnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.688914 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6bdbdac2-5df9-46ab-9965-c5970241fe90" (UID: "6bdbdac2-5df9-46ab-9965-c5970241fe90"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.689173 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bdbdac2-5df9-46ab-9965-c5970241fe90" (UID: "6bdbdac2-5df9-46ab-9965-c5970241fe90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.695930 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6bdbdac2-5df9-46ab-9965-c5970241fe90" (UID: "6bdbdac2-5df9-46ab-9965-c5970241fe90"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.787261 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mvnx\" (UniqueName: \"kubernetes.io/projected/6bdbdac2-5df9-46ab-9965-c5970241fe90-kube-api-access-2mvnx\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.787505 4992 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.787644 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.787719 4992 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6bdbdac2-5df9-46ab-9965-c5970241fe90-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.787800 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bdbdac2-5df9-46ab-9965-c5970241fe90-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:07 crc kubenswrapper[4992]: I1004 04:52:07.788215 4992 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6bdbdac2-5df9-46ab-9965-c5970241fe90-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.482626 4992 generic.go:334] "Generic (PLEG): container finished" podID="9d90c8e7-c146-4390-a587-3bd054f59831" containerID="ec9f8879072b8f14461e3c0430cfefcca2b8986ecb19db42e432adc35960d596" exitCode=0 Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.482704 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" event={"ID":"9d90c8e7-c146-4390-a587-3bd054f59831","Type":"ContainerDied","Data":"ec9f8879072b8f14461e3c0430cfefcca2b8986ecb19db42e432adc35960d596"} Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.483641 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wvk6d" Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.543311 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.545074 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.589677 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wvk6d"] Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.594506 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-wvk6d"] Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.963249 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 04 04:52:08 crc kubenswrapper[4992]: I1004 04:52:08.963314 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 04 04:52:09 crc kubenswrapper[4992]: I1004 04:52:09.496685 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4jk5d" event={"ID":"365fab3e-1695-4e8e-a362-485daebb7f46","Type":"ContainerStarted","Data":"615cc9b13c050a32ab4c27c138dd9d554f693391f24b406b3d14e1d3f48e6cd7"} Oct 04 04:52:10 crc kubenswrapper[4992]: I1004 04:52:10.328878 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bdbdac2-5df9-46ab-9965-c5970241fe90" path="/var/lib/kubelet/pods/6bdbdac2-5df9-46ab-9965-c5970241fe90/volumes" Oct 04 04:52:10 crc kubenswrapper[4992]: I1004 04:52:10.430207 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:10 crc kubenswrapper[4992]: E1004 04:52:10.430636 4992 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:52:10 crc kubenswrapper[4992]: E1004 04:52:10.430716 4992 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:52:10 crc kubenswrapper[4992]: E1004 04:52:10.430835 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift podName:fa875baa-782c-469c-816f-7727c4885d28 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:18.430820245 +0000 UTC m=+1172.278495713 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift") pod "swift-storage-0" (UID: "fa875baa-782c-469c-816f-7727c4885d28") : configmap "swift-ring-files" not found Oct 04 04:52:10 crc kubenswrapper[4992]: I1004 04:52:10.505612 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:10 crc kubenswrapper[4992]: I1004 04:52:10.538975 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-4jk5d" podStartSLOduration=9.538955874 podStartE2EDuration="9.538955874s" podCreationTimestamp="2025-10-04 04:52:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:52:10.536985551 +0000 UTC m=+1164.384661019" watchObservedRunningTime="2025-10-04 04:52:10.538955874 +0000 UTC m=+1164.386631342" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.609941 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.649376 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-ovsdbserver-nb\") pod \"9d90c8e7-c146-4390-a587-3bd054f59831\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.649468 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-config\") pod \"9d90c8e7-c146-4390-a587-3bd054f59831\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.649513 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-dns-svc\") pod \"9d90c8e7-c146-4390-a587-3bd054f59831\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.649553 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9txjb\" (UniqueName: \"kubernetes.io/projected/9d90c8e7-c146-4390-a587-3bd054f59831-kube-api-access-9txjb\") pod \"9d90c8e7-c146-4390-a587-3bd054f59831\" (UID: \"9d90c8e7-c146-4390-a587-3bd054f59831\") " Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.656709 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d90c8e7-c146-4390-a587-3bd054f59831-kube-api-access-9txjb" (OuterVolumeSpecName: "kube-api-access-9txjb") pod "9d90c8e7-c146-4390-a587-3bd054f59831" (UID: "9d90c8e7-c146-4390-a587-3bd054f59831"). InnerVolumeSpecName "kube-api-access-9txjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.695287 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9d90c8e7-c146-4390-a587-3bd054f59831" (UID: "9d90c8e7-c146-4390-a587-3bd054f59831"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.695943 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d90c8e7-c146-4390-a587-3bd054f59831" (UID: "9d90c8e7-c146-4390-a587-3bd054f59831"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.700937 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-config" (OuterVolumeSpecName: "config") pod "9d90c8e7-c146-4390-a587-3bd054f59831" (UID: "9d90c8e7-c146-4390-a587-3bd054f59831"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.752507 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.752552 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.752571 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d90c8e7-c146-4390-a587-3bd054f59831-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:11 crc kubenswrapper[4992]: I1004 04:52:11.752597 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9txjb\" (UniqueName: \"kubernetes.io/projected/9d90c8e7-c146-4390-a587-3bd054f59831-kube-api-access-9txjb\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:12 crc kubenswrapper[4992]: I1004 04:52:12.285546 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:52:12 crc kubenswrapper[4992]: I1004 04:52:12.519637 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" event={"ID":"9d90c8e7-c146-4390-a587-3bd054f59831","Type":"ContainerDied","Data":"148244385d56e5865e8289895da10e97886f4e12d4d06a1bbad07c7222aaac46"} Oct 04 04:52:12 crc kubenswrapper[4992]: I1004 04:52:12.519947 4992 scope.go:117] "RemoveContainer" containerID="ec9f8879072b8f14461e3c0430cfefcca2b8986ecb19db42e432adc35960d596" Oct 04 04:52:12 crc kubenswrapper[4992]: I1004 04:52:12.519689 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-8k5hc" Oct 04 04:52:12 crc kubenswrapper[4992]: I1004 04:52:12.538533 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8k5hc"] Oct 04 04:52:12 crc kubenswrapper[4992]: I1004 04:52:12.544703 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-8k5hc"] Oct 04 04:52:12 crc kubenswrapper[4992]: I1004 04:52:12.549063 4992 scope.go:117] "RemoveContainer" containerID="013ad68102819b8f3c7334c57b21a3fcaea70082cdacc491728bdaed1ccaed0f" Oct 04 04:52:13 crc kubenswrapper[4992]: I1004 04:52:13.531776 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9a31b3de-d7c5-4e1e-8ea8-fce11d130790","Type":"ContainerStarted","Data":"aac29383d0e708e6f9990348fb3915d7fcd62786fc56947ef0f681c5a68f4a84"} Oct 04 04:52:13 crc kubenswrapper[4992]: I1004 04:52:13.532154 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9a31b3de-d7c5-4e1e-8ea8-fce11d130790","Type":"ContainerStarted","Data":"a62f72f6cc003dd51bb873a3f39503fb0e03a5c1bd52c6ad113ac176b065f929"} Oct 04 04:52:13 crc kubenswrapper[4992]: I1004 04:52:13.533283 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 04 04:52:13 crc kubenswrapper[4992]: I1004 04:52:13.561220 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=10.243897634 podStartE2EDuration="17.561197395s" podCreationTimestamp="2025-10-04 04:51:56 +0000 UTC" firstStartedPulling="2025-10-04 04:52:05.753314248 +0000 UTC m=+1159.600989756" lastFinishedPulling="2025-10-04 04:52:13.070614009 +0000 UTC m=+1166.918289517" observedRunningTime="2025-10-04 04:52:13.549218015 +0000 UTC m=+1167.396893483" watchObservedRunningTime="2025-10-04 04:52:13.561197395 +0000 UTC m=+1167.408872863" Oct 04 04:52:14 crc kubenswrapper[4992]: I1004 04:52:14.356089 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d90c8e7-c146-4390-a587-3bd054f59831" path="/var/lib/kubelet/pods/9d90c8e7-c146-4390-a587-3bd054f59831/volumes" Oct 04 04:52:15 crc kubenswrapper[4992]: I1004 04:52:15.133697 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 04 04:52:15 crc kubenswrapper[4992]: I1004 04:52:15.192005 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 04 04:52:16 crc kubenswrapper[4992]: I1004 04:52:16.555875 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9hx4r" event={"ID":"c9af92df-6216-469d-84e9-a09e3b8a648d","Type":"ContainerStarted","Data":"7656c4412b9e6165ae10c6400e5452d93460d8fbd5a993f49d3a18319c9fd8a8"} Oct 04 04:52:16 crc kubenswrapper[4992]: I1004 04:52:16.571737 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9hx4r" podStartSLOduration=2.1757606640000002 podStartE2EDuration="10.571719363s" podCreationTimestamp="2025-10-04 04:52:06 +0000 UTC" firstStartedPulling="2025-10-04 04:52:07.35392634 +0000 UTC m=+1161.201601808" lastFinishedPulling="2025-10-04 04:52:15.749885039 +0000 UTC m=+1169.597560507" observedRunningTime="2025-10-04 04:52:16.570186152 +0000 UTC m=+1170.417861630" watchObservedRunningTime="2025-10-04 04:52:16.571719363 +0000 UTC m=+1170.419394831" Oct 04 04:52:16 crc kubenswrapper[4992]: I1004 04:52:16.767567 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:52:16 crc kubenswrapper[4992]: I1004 04:52:16.824741 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rdf2t"] Oct 04 04:52:16 crc kubenswrapper[4992]: I1004 04:52:16.825043 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" podUID="ec6764f1-efcc-4dac-a681-0af9276accdd" containerName="dnsmasq-dns" containerID="cri-o://5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef" gracePeriod=10 Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.051906 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.133715 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.283713 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.354855 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnw5l\" (UniqueName: \"kubernetes.io/projected/ec6764f1-efcc-4dac-a681-0af9276accdd-kube-api-access-nnw5l\") pod \"ec6764f1-efcc-4dac-a681-0af9276accdd\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.356193 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-dns-svc\") pod \"ec6764f1-efcc-4dac-a681-0af9276accdd\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.356616 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-nb\") pod \"ec6764f1-efcc-4dac-a681-0af9276accdd\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.357343 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-sb\") pod \"ec6764f1-efcc-4dac-a681-0af9276accdd\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.357534 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-config\") pod \"ec6764f1-efcc-4dac-a681-0af9276accdd\" (UID: \"ec6764f1-efcc-4dac-a681-0af9276accdd\") " Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.360670 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec6764f1-efcc-4dac-a681-0af9276accdd-kube-api-access-nnw5l" (OuterVolumeSpecName: "kube-api-access-nnw5l") pod "ec6764f1-efcc-4dac-a681-0af9276accdd" (UID: "ec6764f1-efcc-4dac-a681-0af9276accdd"). InnerVolumeSpecName "kube-api-access-nnw5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.394566 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ec6764f1-efcc-4dac-a681-0af9276accdd" (UID: "ec6764f1-efcc-4dac-a681-0af9276accdd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.396619 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ec6764f1-efcc-4dac-a681-0af9276accdd" (UID: "ec6764f1-efcc-4dac-a681-0af9276accdd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.399820 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-config" (OuterVolumeSpecName: "config") pod "ec6764f1-efcc-4dac-a681-0af9276accdd" (UID: "ec6764f1-efcc-4dac-a681-0af9276accdd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.401065 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec6764f1-efcc-4dac-a681-0af9276accdd" (UID: "ec6764f1-efcc-4dac-a681-0af9276accdd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.459089 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnw5l\" (UniqueName: \"kubernetes.io/projected/ec6764f1-efcc-4dac-a681-0af9276accdd-kube-api-access-nnw5l\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.459123 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.459135 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.459143 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.459153 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec6764f1-efcc-4dac-a681-0af9276accdd-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.563837 4992 generic.go:334] "Generic (PLEG): container finished" podID="ec6764f1-efcc-4dac-a681-0af9276accdd" containerID="5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef" exitCode=0 Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.563871 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.563884 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" event={"ID":"ec6764f1-efcc-4dac-a681-0af9276accdd","Type":"ContainerDied","Data":"5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef"} Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.565243 4992 scope.go:117] "RemoveContainer" containerID="5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.568012 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rdf2t" event={"ID":"ec6764f1-efcc-4dac-a681-0af9276accdd","Type":"ContainerDied","Data":"9ae3a40fe48c645a647b423a01d93dc9434d351b884e981303ae9a1b7d6cb461"} Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.592646 4992 scope.go:117] "RemoveContainer" containerID="c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.618100 4992 scope.go:117] "RemoveContainer" containerID="5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef" Oct 04 04:52:17 crc kubenswrapper[4992]: E1004 04:52:17.618620 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef\": container with ID starting with 5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef not found: ID does not exist" containerID="5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.618664 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef"} err="failed to get container status \"5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef\": rpc error: code = NotFound desc = could not find container \"5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef\": container with ID starting with 5d709864ff914e5070413563a84860f974473de54a24926b2bfdc0f2572d99ef not found: ID does not exist" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.618690 4992 scope.go:117] "RemoveContainer" containerID="c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8" Oct 04 04:52:17 crc kubenswrapper[4992]: E1004 04:52:17.619123 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8\": container with ID starting with c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8 not found: ID does not exist" containerID="c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.619173 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8"} err="failed to get container status \"c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8\": rpc error: code = NotFound desc = could not find container \"c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8\": container with ID starting with c196304824e8d5f89cf28c65ff73b06de8d5fc7a827a4d6c24279b7a94c240b8 not found: ID does not exist" Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.619881 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rdf2t"] Oct 04 04:52:17 crc kubenswrapper[4992]: I1004 04:52:17.626807 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rdf2t"] Oct 04 04:52:18 crc kubenswrapper[4992]: I1004 04:52:18.335234 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec6764f1-efcc-4dac-a681-0af9276accdd" path="/var/lib/kubelet/pods/ec6764f1-efcc-4dac-a681-0af9276accdd/volumes" Oct 04 04:52:18 crc kubenswrapper[4992]: I1004 04:52:18.472015 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:18 crc kubenswrapper[4992]: E1004 04:52:18.472206 4992 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 04:52:18 crc kubenswrapper[4992]: E1004 04:52:18.472232 4992 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 04:52:18 crc kubenswrapper[4992]: E1004 04:52:18.472295 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift podName:fa875baa-782c-469c-816f-7727c4885d28 nodeName:}" failed. No retries permitted until 2025-10-04 04:52:34.472276186 +0000 UTC m=+1188.319951654 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift") pod "swift-storage-0" (UID: "fa875baa-782c-469c-816f-7727c4885d28") : configmap "swift-ring-files" not found Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.190814 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-tqxxk"] Oct 04 04:52:19 crc kubenswrapper[4992]: E1004 04:52:19.191162 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d90c8e7-c146-4390-a587-3bd054f59831" containerName="dnsmasq-dns" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.191175 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d90c8e7-c146-4390-a587-3bd054f59831" containerName="dnsmasq-dns" Oct 04 04:52:19 crc kubenswrapper[4992]: E1004 04:52:19.191186 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec6764f1-efcc-4dac-a681-0af9276accdd" containerName="init" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.191191 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec6764f1-efcc-4dac-a681-0af9276accdd" containerName="init" Oct 04 04:52:19 crc kubenswrapper[4992]: E1004 04:52:19.191222 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec6764f1-efcc-4dac-a681-0af9276accdd" containerName="dnsmasq-dns" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.191242 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec6764f1-efcc-4dac-a681-0af9276accdd" containerName="dnsmasq-dns" Oct 04 04:52:19 crc kubenswrapper[4992]: E1004 04:52:19.191253 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d90c8e7-c146-4390-a587-3bd054f59831" containerName="init" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.191258 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d90c8e7-c146-4390-a587-3bd054f59831" containerName="init" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.191432 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d90c8e7-c146-4390-a587-3bd054f59831" containerName="dnsmasq-dns" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.191452 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec6764f1-efcc-4dac-a681-0af9276accdd" containerName="dnsmasq-dns" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.191937 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tqxxk" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.197251 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-tqxxk"] Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.286109 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8mj7\" (UniqueName: \"kubernetes.io/projected/3365eb93-bd5e-4dee-9386-1382ee1c952e-kube-api-access-g8mj7\") pod \"keystone-db-create-tqxxk\" (UID: \"3365eb93-bd5e-4dee-9386-1382ee1c952e\") " pod="openstack/keystone-db-create-tqxxk" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.372921 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-4c65b"] Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.373885 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4c65b" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.383942 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-4c65b"] Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.387308 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bdtg\" (UniqueName: \"kubernetes.io/projected/a21eec04-f852-4e59-ba65-c40a22393706-kube-api-access-7bdtg\") pod \"placement-db-create-4c65b\" (UID: \"a21eec04-f852-4e59-ba65-c40a22393706\") " pod="openstack/placement-db-create-4c65b" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.387573 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8mj7\" (UniqueName: \"kubernetes.io/projected/3365eb93-bd5e-4dee-9386-1382ee1c952e-kube-api-access-g8mj7\") pod \"keystone-db-create-tqxxk\" (UID: \"3365eb93-bd5e-4dee-9386-1382ee1c952e\") " pod="openstack/keystone-db-create-tqxxk" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.414644 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8mj7\" (UniqueName: \"kubernetes.io/projected/3365eb93-bd5e-4dee-9386-1382ee1c952e-kube-api-access-g8mj7\") pod \"keystone-db-create-tqxxk\" (UID: \"3365eb93-bd5e-4dee-9386-1382ee1c952e\") " pod="openstack/keystone-db-create-tqxxk" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.489289 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bdtg\" (UniqueName: \"kubernetes.io/projected/a21eec04-f852-4e59-ba65-c40a22393706-kube-api-access-7bdtg\") pod \"placement-db-create-4c65b\" (UID: \"a21eec04-f852-4e59-ba65-c40a22393706\") " pod="openstack/placement-db-create-4c65b" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.508716 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bdtg\" (UniqueName: \"kubernetes.io/projected/a21eec04-f852-4e59-ba65-c40a22393706-kube-api-access-7bdtg\") pod \"placement-db-create-4c65b\" (UID: \"a21eec04-f852-4e59-ba65-c40a22393706\") " pod="openstack/placement-db-create-4c65b" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.515052 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tqxxk" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.596842 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-f2flr"] Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.597745 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f2flr" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.608630 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-f2flr"] Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.687028 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4c65b" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.693535 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbbnk\" (UniqueName: \"kubernetes.io/projected/dedeadd4-7b3f-433c-b310-be8f83cbce0b-kube-api-access-xbbnk\") pod \"glance-db-create-f2flr\" (UID: \"dedeadd4-7b3f-433c-b310-be8f83cbce0b\") " pod="openstack/glance-db-create-f2flr" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.798049 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbbnk\" (UniqueName: \"kubernetes.io/projected/dedeadd4-7b3f-433c-b310-be8f83cbce0b-kube-api-access-xbbnk\") pod \"glance-db-create-f2flr\" (UID: \"dedeadd4-7b3f-433c-b310-be8f83cbce0b\") " pod="openstack/glance-db-create-f2flr" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.817159 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbbnk\" (UniqueName: \"kubernetes.io/projected/dedeadd4-7b3f-433c-b310-be8f83cbce0b-kube-api-access-xbbnk\") pod \"glance-db-create-f2flr\" (UID: \"dedeadd4-7b3f-433c-b310-be8f83cbce0b\") " pod="openstack/glance-db-create-f2flr" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.953433 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f2flr" Oct 04 04:52:19 crc kubenswrapper[4992]: I1004 04:52:19.977342 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-tqxxk"] Oct 04 04:52:20 crc kubenswrapper[4992]: W1004 04:52:20.106586 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda21eec04_f852_4e59_ba65_c40a22393706.slice/crio-6eae0837ea55cf266f16bd0047a119b01c2af2ba1e0b76bf38191673577c61b8 WatchSource:0}: Error finding container 6eae0837ea55cf266f16bd0047a119b01c2af2ba1e0b76bf38191673577c61b8: Status 404 returned error can't find the container with id 6eae0837ea55cf266f16bd0047a119b01c2af2ba1e0b76bf38191673577c61b8 Oct 04 04:52:20 crc kubenswrapper[4992]: I1004 04:52:20.112833 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-4c65b"] Oct 04 04:52:20 crc kubenswrapper[4992]: I1004 04:52:20.387581 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-f2flr"] Oct 04 04:52:20 crc kubenswrapper[4992]: I1004 04:52:20.600501 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4c65b" event={"ID":"a21eec04-f852-4e59-ba65-c40a22393706","Type":"ContainerStarted","Data":"6eae0837ea55cf266f16bd0047a119b01c2af2ba1e0b76bf38191673577c61b8"} Oct 04 04:52:20 crc kubenswrapper[4992]: I1004 04:52:20.601307 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tqxxk" event={"ID":"3365eb93-bd5e-4dee-9386-1382ee1c952e","Type":"ContainerStarted","Data":"5dd116e68d0f9ca7e157c87c1becfe9161e9af8543455830e4b8a66f7d6df336"} Oct 04 04:52:20 crc kubenswrapper[4992]: I1004 04:52:20.603681 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f2flr" event={"ID":"dedeadd4-7b3f-433c-b310-be8f83cbce0b","Type":"ContainerStarted","Data":"845097a082f7165c763f0a11dbb9114b0426ff6c691fe083f651fbeabb34db5e"} Oct 04 04:52:20 crc kubenswrapper[4992]: I1004 04:52:20.774569 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8m2tq" podUID="06640053-739f-4a61-84e0-c6ed4876a249" containerName="ovn-controller" probeResult="failure" output=< Oct 04 04:52:20 crc kubenswrapper[4992]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 04:52:20 crc kubenswrapper[4992]: > Oct 04 04:52:21 crc kubenswrapper[4992]: I1004 04:52:21.614325 4992 generic.go:334] "Generic (PLEG): container finished" podID="a21eec04-f852-4e59-ba65-c40a22393706" containerID="dcd3dbfe58eb58206f09ba1f68da07606ec29892fa372bcf976bda70525664e2" exitCode=0 Oct 04 04:52:21 crc kubenswrapper[4992]: I1004 04:52:21.614572 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4c65b" event={"ID":"a21eec04-f852-4e59-ba65-c40a22393706","Type":"ContainerDied","Data":"dcd3dbfe58eb58206f09ba1f68da07606ec29892fa372bcf976bda70525664e2"} Oct 04 04:52:21 crc kubenswrapper[4992]: I1004 04:52:21.616838 4992 generic.go:334] "Generic (PLEG): container finished" podID="3365eb93-bd5e-4dee-9386-1382ee1c952e" containerID="f5ef9c36afc2be9c54a5499cc20ebbfb57ef5a929e3b1842919345cf22c80c9a" exitCode=0 Oct 04 04:52:21 crc kubenswrapper[4992]: I1004 04:52:21.616923 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tqxxk" event={"ID":"3365eb93-bd5e-4dee-9386-1382ee1c952e","Type":"ContainerDied","Data":"f5ef9c36afc2be9c54a5499cc20ebbfb57ef5a929e3b1842919345cf22c80c9a"} Oct 04 04:52:21 crc kubenswrapper[4992]: I1004 04:52:21.619215 4992 generic.go:334] "Generic (PLEG): container finished" podID="dedeadd4-7b3f-433c-b310-be8f83cbce0b" containerID="f76dd8f975bddf7ce304c807b17f9b4bb5e399ed4439bf85f198ba1476473c98" exitCode=0 Oct 04 04:52:21 crc kubenswrapper[4992]: I1004 04:52:21.619262 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f2flr" event={"ID":"dedeadd4-7b3f-433c-b310-be8f83cbce0b","Type":"ContainerDied","Data":"f76dd8f975bddf7ce304c807b17f9b4bb5e399ed4439bf85f198ba1476473c98"} Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.075127 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tqxxk" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.082894 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4c65b" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.088832 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f2flr" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.250630 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8mj7\" (UniqueName: \"kubernetes.io/projected/3365eb93-bd5e-4dee-9386-1382ee1c952e-kube-api-access-g8mj7\") pod \"3365eb93-bd5e-4dee-9386-1382ee1c952e\" (UID: \"3365eb93-bd5e-4dee-9386-1382ee1c952e\") " Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.250712 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbbnk\" (UniqueName: \"kubernetes.io/projected/dedeadd4-7b3f-433c-b310-be8f83cbce0b-kube-api-access-xbbnk\") pod \"dedeadd4-7b3f-433c-b310-be8f83cbce0b\" (UID: \"dedeadd4-7b3f-433c-b310-be8f83cbce0b\") " Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.250804 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bdtg\" (UniqueName: \"kubernetes.io/projected/a21eec04-f852-4e59-ba65-c40a22393706-kube-api-access-7bdtg\") pod \"a21eec04-f852-4e59-ba65-c40a22393706\" (UID: \"a21eec04-f852-4e59-ba65-c40a22393706\") " Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.256956 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3365eb93-bd5e-4dee-9386-1382ee1c952e-kube-api-access-g8mj7" (OuterVolumeSpecName: "kube-api-access-g8mj7") pod "3365eb93-bd5e-4dee-9386-1382ee1c952e" (UID: "3365eb93-bd5e-4dee-9386-1382ee1c952e"). InnerVolumeSpecName "kube-api-access-g8mj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.257594 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dedeadd4-7b3f-433c-b310-be8f83cbce0b-kube-api-access-xbbnk" (OuterVolumeSpecName: "kube-api-access-xbbnk") pod "dedeadd4-7b3f-433c-b310-be8f83cbce0b" (UID: "dedeadd4-7b3f-433c-b310-be8f83cbce0b"). InnerVolumeSpecName "kube-api-access-xbbnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.260080 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a21eec04-f852-4e59-ba65-c40a22393706-kube-api-access-7bdtg" (OuterVolumeSpecName: "kube-api-access-7bdtg") pod "a21eec04-f852-4e59-ba65-c40a22393706" (UID: "a21eec04-f852-4e59-ba65-c40a22393706"). InnerVolumeSpecName "kube-api-access-7bdtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.352151 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bdtg\" (UniqueName: \"kubernetes.io/projected/a21eec04-f852-4e59-ba65-c40a22393706-kube-api-access-7bdtg\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.352179 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8mj7\" (UniqueName: \"kubernetes.io/projected/3365eb93-bd5e-4dee-9386-1382ee1c952e-kube-api-access-g8mj7\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.352188 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbbnk\" (UniqueName: \"kubernetes.io/projected/dedeadd4-7b3f-433c-b310-be8f83cbce0b-kube-api-access-xbbnk\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.635149 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tqxxk" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.635142 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tqxxk" event={"ID":"3365eb93-bd5e-4dee-9386-1382ee1c952e","Type":"ContainerDied","Data":"5dd116e68d0f9ca7e157c87c1becfe9161e9af8543455830e4b8a66f7d6df336"} Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.635485 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dd116e68d0f9ca7e157c87c1becfe9161e9af8543455830e4b8a66f7d6df336" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.637786 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f2flr" event={"ID":"dedeadd4-7b3f-433c-b310-be8f83cbce0b","Type":"ContainerDied","Data":"845097a082f7165c763f0a11dbb9114b0426ff6c691fe083f651fbeabb34db5e"} Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.637811 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f2flr" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.637825 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="845097a082f7165c763f0a11dbb9114b0426ff6c691fe083f651fbeabb34db5e" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.639713 4992 generic.go:334] "Generic (PLEG): container finished" podID="c9af92df-6216-469d-84e9-a09e3b8a648d" containerID="7656c4412b9e6165ae10c6400e5452d93460d8fbd5a993f49d3a18319c9fd8a8" exitCode=0 Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.639764 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9hx4r" event={"ID":"c9af92df-6216-469d-84e9-a09e3b8a648d","Type":"ContainerDied","Data":"7656c4412b9e6165ae10c6400e5452d93460d8fbd5a993f49d3a18319c9fd8a8"} Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.642007 4992 generic.go:334] "Generic (PLEG): container finished" podID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerID="c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44" exitCode=0 Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.642130 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52","Type":"ContainerDied","Data":"c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44"} Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.648880 4992 generic.go:334] "Generic (PLEG): container finished" podID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerID="f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd" exitCode=0 Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.649105 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0c7fa3bc-a429-4f65-aac7-a08293b802eb","Type":"ContainerDied","Data":"f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd"} Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.652847 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-4c65b" event={"ID":"a21eec04-f852-4e59-ba65-c40a22393706","Type":"ContainerDied","Data":"6eae0837ea55cf266f16bd0047a119b01c2af2ba1e0b76bf38191673577c61b8"} Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.652919 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6eae0837ea55cf266f16bd0047a119b01c2af2ba1e0b76bf38191673577c61b8" Oct 04 04:52:23 crc kubenswrapper[4992]: I1004 04:52:23.653068 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-4c65b" Oct 04 04:52:24 crc kubenswrapper[4992]: I1004 04:52:24.672638 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52","Type":"ContainerStarted","Data":"328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049"} Oct 04 04:52:24 crc kubenswrapper[4992]: I1004 04:52:24.673469 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:52:24 crc kubenswrapper[4992]: I1004 04:52:24.674786 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0c7fa3bc-a429-4f65-aac7-a08293b802eb","Type":"ContainerStarted","Data":"fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15"} Oct 04 04:52:24 crc kubenswrapper[4992]: I1004 04:52:24.674972 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 04:52:24 crc kubenswrapper[4992]: I1004 04:52:24.704621 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.326294642 podStartE2EDuration="1m0.704599302s" podCreationTimestamp="2025-10-04 04:51:24 +0000 UTC" firstStartedPulling="2025-10-04 04:51:38.253601751 +0000 UTC m=+1132.101277219" lastFinishedPulling="2025-10-04 04:51:48.631906411 +0000 UTC m=+1142.479581879" observedRunningTime="2025-10-04 04:52:24.701181031 +0000 UTC m=+1178.548856539" watchObservedRunningTime="2025-10-04 04:52:24.704599302 +0000 UTC m=+1178.552274770" Oct 04 04:52:24 crc kubenswrapper[4992]: I1004 04:52:24.728009 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=51.740357508 podStartE2EDuration="1m0.727977305s" podCreationTimestamp="2025-10-04 04:51:24 +0000 UTC" firstStartedPulling="2025-10-04 04:51:37.951431468 +0000 UTC m=+1131.799106946" lastFinishedPulling="2025-10-04 04:51:46.939051275 +0000 UTC m=+1140.786726743" observedRunningTime="2025-10-04 04:52:24.7240432 +0000 UTC m=+1178.571718738" watchObservedRunningTime="2025-10-04 04:52:24.727977305 +0000 UTC m=+1178.575652813" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.024499 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.196895 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-ring-data-devices\") pod \"c9af92df-6216-469d-84e9-a09e3b8a648d\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.196990 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-combined-ca-bundle\") pod \"c9af92df-6216-469d-84e9-a09e3b8a648d\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.197073 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-scripts\") pod \"c9af92df-6216-469d-84e9-a09e3b8a648d\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.197108 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-swiftconf\") pod \"c9af92df-6216-469d-84e9-a09e3b8a648d\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.197135 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c9af92df-6216-469d-84e9-a09e3b8a648d-etc-swift\") pod \"c9af92df-6216-469d-84e9-a09e3b8a648d\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.197169 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-dispersionconf\") pod \"c9af92df-6216-469d-84e9-a09e3b8a648d\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.197206 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h4zk\" (UniqueName: \"kubernetes.io/projected/c9af92df-6216-469d-84e9-a09e3b8a648d-kube-api-access-2h4zk\") pod \"c9af92df-6216-469d-84e9-a09e3b8a648d\" (UID: \"c9af92df-6216-469d-84e9-a09e3b8a648d\") " Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.197332 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c9af92df-6216-469d-84e9-a09e3b8a648d" (UID: "c9af92df-6216-469d-84e9-a09e3b8a648d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.197552 4992 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.198200 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9af92df-6216-469d-84e9-a09e3b8a648d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c9af92df-6216-469d-84e9-a09e3b8a648d" (UID: "c9af92df-6216-469d-84e9-a09e3b8a648d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.201992 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9af92df-6216-469d-84e9-a09e3b8a648d-kube-api-access-2h4zk" (OuterVolumeSpecName: "kube-api-access-2h4zk") pod "c9af92df-6216-469d-84e9-a09e3b8a648d" (UID: "c9af92df-6216-469d-84e9-a09e3b8a648d"). InnerVolumeSpecName "kube-api-access-2h4zk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.211197 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c9af92df-6216-469d-84e9-a09e3b8a648d" (UID: "c9af92df-6216-469d-84e9-a09e3b8a648d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.220274 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-scripts" (OuterVolumeSpecName: "scripts") pod "c9af92df-6216-469d-84e9-a09e3b8a648d" (UID: "c9af92df-6216-469d-84e9-a09e3b8a648d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.225504 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c9af92df-6216-469d-84e9-a09e3b8a648d" (UID: "c9af92df-6216-469d-84e9-a09e3b8a648d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.245652 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9af92df-6216-469d-84e9-a09e3b8a648d" (UID: "c9af92df-6216-469d-84e9-a09e3b8a648d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.299697 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c9af92df-6216-469d-84e9-a09e3b8a648d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.299732 4992 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.299753 4992 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c9af92df-6216-469d-84e9-a09e3b8a648d-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.299764 4992 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.299778 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h4zk\" (UniqueName: \"kubernetes.io/projected/c9af92df-6216-469d-84e9-a09e3b8a648d-kube-api-access-2h4zk\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.299791 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9af92df-6216-469d-84e9-a09e3b8a648d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.682825 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9hx4r" event={"ID":"c9af92df-6216-469d-84e9-a09e3b8a648d","Type":"ContainerDied","Data":"7c703c998c8be3fd65f5f59ebd45228e169257833266463910de7b176d7feb2a"} Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.682891 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c703c998c8be3fd65f5f59ebd45228e169257833266463910de7b176d7feb2a" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.682846 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9hx4r" Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.788768 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8m2tq" podUID="06640053-739f-4a61-84e0-c6ed4876a249" containerName="ovn-controller" probeResult="failure" output=< Oct 04 04:52:25 crc kubenswrapper[4992]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 04:52:25 crc kubenswrapper[4992]: > Oct 04 04:52:25 crc kubenswrapper[4992]: I1004 04:52:25.803249 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:52:27 crc kubenswrapper[4992]: I1004 04:52:27.165978 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.212767 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e810-account-create-bfskd"] Oct 04 04:52:29 crc kubenswrapper[4992]: E1004 04:52:29.213726 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3365eb93-bd5e-4dee-9386-1382ee1c952e" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.213744 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3365eb93-bd5e-4dee-9386-1382ee1c952e" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: E1004 04:52:29.213764 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dedeadd4-7b3f-433c-b310-be8f83cbce0b" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.213773 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="dedeadd4-7b3f-433c-b310-be8f83cbce0b" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: E1004 04:52:29.213792 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a21eec04-f852-4e59-ba65-c40a22393706" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.213801 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a21eec04-f852-4e59-ba65-c40a22393706" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: E1004 04:52:29.213813 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9af92df-6216-469d-84e9-a09e3b8a648d" containerName="swift-ring-rebalance" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.213822 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9af92df-6216-469d-84e9-a09e3b8a648d" containerName="swift-ring-rebalance" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.214026 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3365eb93-bd5e-4dee-9386-1382ee1c952e" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.214048 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="dedeadd4-7b3f-433c-b310-be8f83cbce0b" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.214066 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9af92df-6216-469d-84e9-a09e3b8a648d" containerName="swift-ring-rebalance" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.214080 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="a21eec04-f852-4e59-ba65-c40a22393706" containerName="mariadb-database-create" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.214837 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e810-account-create-bfskd" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.216702 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.229395 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e810-account-create-bfskd"] Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.366331 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrjns\" (UniqueName: \"kubernetes.io/projected/81a290cb-94bf-41ad-87bf-ccc8a713e03d-kube-api-access-lrjns\") pod \"keystone-e810-account-create-bfskd\" (UID: \"81a290cb-94bf-41ad-87bf-ccc8a713e03d\") " pod="openstack/keystone-e810-account-create-bfskd" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.424642 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9169-account-create-dpkj4"] Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.425716 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9169-account-create-dpkj4" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.439277 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9169-account-create-dpkj4"] Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.439702 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.467477 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrjns\" (UniqueName: \"kubernetes.io/projected/81a290cb-94bf-41ad-87bf-ccc8a713e03d-kube-api-access-lrjns\") pod \"keystone-e810-account-create-bfskd\" (UID: \"81a290cb-94bf-41ad-87bf-ccc8a713e03d\") " pod="openstack/keystone-e810-account-create-bfskd" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.485683 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrjns\" (UniqueName: \"kubernetes.io/projected/81a290cb-94bf-41ad-87bf-ccc8a713e03d-kube-api-access-lrjns\") pod \"keystone-e810-account-create-bfskd\" (UID: \"81a290cb-94bf-41ad-87bf-ccc8a713e03d\") " pod="openstack/keystone-e810-account-create-bfskd" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.546212 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e810-account-create-bfskd" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.569387 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvp8c\" (UniqueName: \"kubernetes.io/projected/dd465ca0-c12b-4f3c-8de2-377ac6d81f89-kube-api-access-tvp8c\") pod \"placement-9169-account-create-dpkj4\" (UID: \"dd465ca0-c12b-4f3c-8de2-377ac6d81f89\") " pod="openstack/placement-9169-account-create-dpkj4" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.670981 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvp8c\" (UniqueName: \"kubernetes.io/projected/dd465ca0-c12b-4f3c-8de2-377ac6d81f89-kube-api-access-tvp8c\") pod \"placement-9169-account-create-dpkj4\" (UID: \"dd465ca0-c12b-4f3c-8de2-377ac6d81f89\") " pod="openstack/placement-9169-account-create-dpkj4" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.691170 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvp8c\" (UniqueName: \"kubernetes.io/projected/dd465ca0-c12b-4f3c-8de2-377ac6d81f89-kube-api-access-tvp8c\") pod \"placement-9169-account-create-dpkj4\" (UID: \"dd465ca0-c12b-4f3c-8de2-377ac6d81f89\") " pod="openstack/placement-9169-account-create-dpkj4" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.741078 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9169-account-create-dpkj4" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.775801 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-eff6-account-create-sz54v"] Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.778328 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eff6-account-create-sz54v" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.780703 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.799926 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-eff6-account-create-sz54v"] Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.876242 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c59bg\" (UniqueName: \"kubernetes.io/projected/2b8c7673-09e9-4b88-9970-5a3a01eac6a3-kube-api-access-c59bg\") pod \"glance-eff6-account-create-sz54v\" (UID: \"2b8c7673-09e9-4b88-9970-5a3a01eac6a3\") " pod="openstack/glance-eff6-account-create-sz54v" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.978027 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c59bg\" (UniqueName: \"kubernetes.io/projected/2b8c7673-09e9-4b88-9970-5a3a01eac6a3-kube-api-access-c59bg\") pod \"glance-eff6-account-create-sz54v\" (UID: \"2b8c7673-09e9-4b88-9970-5a3a01eac6a3\") " pod="openstack/glance-eff6-account-create-sz54v" Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.991687 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e810-account-create-bfskd"] Oct 04 04:52:29 crc kubenswrapper[4992]: I1004 04:52:29.998093 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c59bg\" (UniqueName: \"kubernetes.io/projected/2b8c7673-09e9-4b88-9970-5a3a01eac6a3-kube-api-access-c59bg\") pod \"glance-eff6-account-create-sz54v\" (UID: \"2b8c7673-09e9-4b88-9970-5a3a01eac6a3\") " pod="openstack/glance-eff6-account-create-sz54v" Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.111159 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eff6-account-create-sz54v" Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.185243 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9169-account-create-dpkj4"] Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.586768 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-eff6-account-create-sz54v"] Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.726608 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eff6-account-create-sz54v" event={"ID":"2b8c7673-09e9-4b88-9970-5a3a01eac6a3","Type":"ContainerStarted","Data":"403c2dcd71d60a9a2617a4a0927176fe36d8ac1f0bb00a658036b2b1a92fc444"} Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.728348 4992 generic.go:334] "Generic (PLEG): container finished" podID="dd465ca0-c12b-4f3c-8de2-377ac6d81f89" containerID="4419ca98afc2b0d3056a2de9cd37f6807c07a3755f066da6d965a09614690f53" exitCode=0 Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.728413 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9169-account-create-dpkj4" event={"ID":"dd465ca0-c12b-4f3c-8de2-377ac6d81f89","Type":"ContainerDied","Data":"4419ca98afc2b0d3056a2de9cd37f6807c07a3755f066da6d965a09614690f53"} Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.728467 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9169-account-create-dpkj4" event={"ID":"dd465ca0-c12b-4f3c-8de2-377ac6d81f89","Type":"ContainerStarted","Data":"7b87959330262b167cfc97ea651c4760128940ab43c7d4481743bf2db1cffd3e"} Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.729978 4992 generic.go:334] "Generic (PLEG): container finished" podID="81a290cb-94bf-41ad-87bf-ccc8a713e03d" containerID="96c86b7166336c91921d9a11958638b6d887892378a8b31751fb5e321561b55a" exitCode=0 Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.730021 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e810-account-create-bfskd" event={"ID":"81a290cb-94bf-41ad-87bf-ccc8a713e03d","Type":"ContainerDied","Data":"96c86b7166336c91921d9a11958638b6d887892378a8b31751fb5e321561b55a"} Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.730039 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e810-account-create-bfskd" event={"ID":"81a290cb-94bf-41ad-87bf-ccc8a713e03d","Type":"ContainerStarted","Data":"fa3b31fc560809b47bc8abafe1cfb769a427068b61dd4e42fee34c9a016eb94f"} Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.787043 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-8m2tq" podUID="06640053-739f-4a61-84e0-c6ed4876a249" containerName="ovn-controller" probeResult="failure" output=< Oct 04 04:52:30 crc kubenswrapper[4992]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 04:52:30 crc kubenswrapper[4992]: > Oct 04 04:52:30 crc kubenswrapper[4992]: I1004 04:52:30.801671 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-qv6w9" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.000469 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8m2tq-config-469c5"] Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.001802 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.004002 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.015241 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8m2tq-config-469c5"] Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.097641 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run-ovn\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.097923 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.098004 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-scripts\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.098059 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-log-ovn\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.098161 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf2st\" (UniqueName: \"kubernetes.io/projected/9ef0036a-de1e-40ee-99b3-5d15b2775955-kube-api-access-xf2st\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.098376 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-additional-scripts\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.199880 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run-ovn\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.199993 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.200020 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-scripts\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.200045 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-log-ovn\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.200089 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf2st\" (UniqueName: \"kubernetes.io/projected/9ef0036a-de1e-40ee-99b3-5d15b2775955-kube-api-access-xf2st\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.200122 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run-ovn\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.200142 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.200127 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-additional-scripts\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.200210 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-log-ovn\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.201043 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-additional-scripts\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.202495 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-scripts\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.217917 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf2st\" (UniqueName: \"kubernetes.io/projected/9ef0036a-de1e-40ee-99b3-5d15b2775955-kube-api-access-xf2st\") pod \"ovn-controller-8m2tq-config-469c5\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.320725 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.744651 4992 generic.go:334] "Generic (PLEG): container finished" podID="2b8c7673-09e9-4b88-9970-5a3a01eac6a3" containerID="16c3322b8796d7b6a6fba9e2d760ce5c71fba07e153ce98649e53d312d65c58c" exitCode=0 Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.744788 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eff6-account-create-sz54v" event={"ID":"2b8c7673-09e9-4b88-9970-5a3a01eac6a3","Type":"ContainerDied","Data":"16c3322b8796d7b6a6fba9e2d760ce5c71fba07e153ce98649e53d312d65c58c"} Oct 04 04:52:31 crc kubenswrapper[4992]: I1004 04:52:31.758842 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8m2tq-config-469c5"] Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.118037 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9169-account-create-dpkj4" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.130481 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e810-account-create-bfskd" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.226884 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvp8c\" (UniqueName: \"kubernetes.io/projected/dd465ca0-c12b-4f3c-8de2-377ac6d81f89-kube-api-access-tvp8c\") pod \"dd465ca0-c12b-4f3c-8de2-377ac6d81f89\" (UID: \"dd465ca0-c12b-4f3c-8de2-377ac6d81f89\") " Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.226930 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrjns\" (UniqueName: \"kubernetes.io/projected/81a290cb-94bf-41ad-87bf-ccc8a713e03d-kube-api-access-lrjns\") pod \"81a290cb-94bf-41ad-87bf-ccc8a713e03d\" (UID: \"81a290cb-94bf-41ad-87bf-ccc8a713e03d\") " Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.232481 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81a290cb-94bf-41ad-87bf-ccc8a713e03d-kube-api-access-lrjns" (OuterVolumeSpecName: "kube-api-access-lrjns") pod "81a290cb-94bf-41ad-87bf-ccc8a713e03d" (UID: "81a290cb-94bf-41ad-87bf-ccc8a713e03d"). InnerVolumeSpecName "kube-api-access-lrjns". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.233119 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd465ca0-c12b-4f3c-8de2-377ac6d81f89-kube-api-access-tvp8c" (OuterVolumeSpecName: "kube-api-access-tvp8c") pod "dd465ca0-c12b-4f3c-8de2-377ac6d81f89" (UID: "dd465ca0-c12b-4f3c-8de2-377ac6d81f89"). InnerVolumeSpecName "kube-api-access-tvp8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.328321 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvp8c\" (UniqueName: \"kubernetes.io/projected/dd465ca0-c12b-4f3c-8de2-377ac6d81f89-kube-api-access-tvp8c\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.328381 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrjns\" (UniqueName: \"kubernetes.io/projected/81a290cb-94bf-41ad-87bf-ccc8a713e03d-kube-api-access-lrjns\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.753804 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e810-account-create-bfskd" event={"ID":"81a290cb-94bf-41ad-87bf-ccc8a713e03d","Type":"ContainerDied","Data":"fa3b31fc560809b47bc8abafe1cfb769a427068b61dd4e42fee34c9a016eb94f"} Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.754129 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa3b31fc560809b47bc8abafe1cfb769a427068b61dd4e42fee34c9a016eb94f" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.753815 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e810-account-create-bfskd" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.755940 4992 generic.go:334] "Generic (PLEG): container finished" podID="9ef0036a-de1e-40ee-99b3-5d15b2775955" containerID="b331148fadb0665defb0faacc69aa53d565b349c47baf03f4e851a44fe50069d" exitCode=0 Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.756008 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq-config-469c5" event={"ID":"9ef0036a-de1e-40ee-99b3-5d15b2775955","Type":"ContainerDied","Data":"b331148fadb0665defb0faacc69aa53d565b349c47baf03f4e851a44fe50069d"} Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.756033 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq-config-469c5" event={"ID":"9ef0036a-de1e-40ee-99b3-5d15b2775955","Type":"ContainerStarted","Data":"f3bc1c2abb73173cd51ccabcc2520e922251d2a2bd66fa9ba91433e45df40756"} Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.758651 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9169-account-create-dpkj4" event={"ID":"dd465ca0-c12b-4f3c-8de2-377ac6d81f89","Type":"ContainerDied","Data":"7b87959330262b167cfc97ea651c4760128940ab43c7d4481743bf2db1cffd3e"} Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.758701 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b87959330262b167cfc97ea651c4760128940ab43c7d4481743bf2db1cffd3e" Oct 04 04:52:32 crc kubenswrapper[4992]: I1004 04:52:32.758667 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9169-account-create-dpkj4" Oct 04 04:52:33 crc kubenswrapper[4992]: I1004 04:52:33.142065 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eff6-account-create-sz54v" Oct 04 04:52:33 crc kubenswrapper[4992]: I1004 04:52:33.241271 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c59bg\" (UniqueName: \"kubernetes.io/projected/2b8c7673-09e9-4b88-9970-5a3a01eac6a3-kube-api-access-c59bg\") pod \"2b8c7673-09e9-4b88-9970-5a3a01eac6a3\" (UID: \"2b8c7673-09e9-4b88-9970-5a3a01eac6a3\") " Oct 04 04:52:33 crc kubenswrapper[4992]: I1004 04:52:33.246389 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b8c7673-09e9-4b88-9970-5a3a01eac6a3-kube-api-access-c59bg" (OuterVolumeSpecName: "kube-api-access-c59bg") pod "2b8c7673-09e9-4b88-9970-5a3a01eac6a3" (UID: "2b8c7673-09e9-4b88-9970-5a3a01eac6a3"). InnerVolumeSpecName "kube-api-access-c59bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:33 crc kubenswrapper[4992]: I1004 04:52:33.343226 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c59bg\" (UniqueName: \"kubernetes.io/projected/2b8c7673-09e9-4b88-9970-5a3a01eac6a3-kube-api-access-c59bg\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:33 crc kubenswrapper[4992]: I1004 04:52:33.768864 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eff6-account-create-sz54v" Oct 04 04:52:33 crc kubenswrapper[4992]: I1004 04:52:33.768859 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eff6-account-create-sz54v" event={"ID":"2b8c7673-09e9-4b88-9970-5a3a01eac6a3","Type":"ContainerDied","Data":"403c2dcd71d60a9a2617a4a0927176fe36d8ac1f0bb00a658036b2b1a92fc444"} Oct 04 04:52:33 crc kubenswrapper[4992]: I1004 04:52:33.769078 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="403c2dcd71d60a9a2617a4a0927176fe36d8ac1f0bb00a658036b2b1a92fc444" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.077661 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261071 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-scripts\") pod \"9ef0036a-de1e-40ee-99b3-5d15b2775955\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261128 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-log-ovn\") pod \"9ef0036a-de1e-40ee-99b3-5d15b2775955\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261181 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf2st\" (UniqueName: \"kubernetes.io/projected/9ef0036a-de1e-40ee-99b3-5d15b2775955-kube-api-access-xf2st\") pod \"9ef0036a-de1e-40ee-99b3-5d15b2775955\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261227 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9ef0036a-de1e-40ee-99b3-5d15b2775955" (UID: "9ef0036a-de1e-40ee-99b3-5d15b2775955"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261244 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-additional-scripts\") pod \"9ef0036a-de1e-40ee-99b3-5d15b2775955\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261290 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run-ovn\") pod \"9ef0036a-de1e-40ee-99b3-5d15b2775955\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261333 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run\") pod \"9ef0036a-de1e-40ee-99b3-5d15b2775955\" (UID: \"9ef0036a-de1e-40ee-99b3-5d15b2775955\") " Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261456 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run" (OuterVolumeSpecName: "var-run") pod "9ef0036a-de1e-40ee-99b3-5d15b2775955" (UID: "9ef0036a-de1e-40ee-99b3-5d15b2775955"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261554 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9ef0036a-de1e-40ee-99b3-5d15b2775955" (UID: "9ef0036a-de1e-40ee-99b3-5d15b2775955"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261783 4992 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261803 4992 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.261817 4992 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ef0036a-de1e-40ee-99b3-5d15b2775955-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.262100 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9ef0036a-de1e-40ee-99b3-5d15b2775955" (UID: "9ef0036a-de1e-40ee-99b3-5d15b2775955"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.262386 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-scripts" (OuterVolumeSpecName: "scripts") pod "9ef0036a-de1e-40ee-99b3-5d15b2775955" (UID: "9ef0036a-de1e-40ee-99b3-5d15b2775955"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.265688 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ef0036a-de1e-40ee-99b3-5d15b2775955-kube-api-access-xf2st" (OuterVolumeSpecName: "kube-api-access-xf2st") pod "9ef0036a-de1e-40ee-99b3-5d15b2775955" (UID: "9ef0036a-de1e-40ee-99b3-5d15b2775955"). InnerVolumeSpecName "kube-api-access-xf2st". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.363927 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.363971 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf2st\" (UniqueName: \"kubernetes.io/projected/9ef0036a-de1e-40ee-99b3-5d15b2775955-kube-api-access-xf2st\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.363985 4992 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ef0036a-de1e-40ee-99b3-5d15b2775955-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.567415 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.572766 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fa875baa-782c-469c-816f-7727c4885d28-etc-swift\") pod \"swift-storage-0\" (UID: \"fa875baa-782c-469c-816f-7727c4885d28\") " pod="openstack/swift-storage-0" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.729466 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.779493 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq-config-469c5" event={"ID":"9ef0036a-de1e-40ee-99b3-5d15b2775955","Type":"ContainerDied","Data":"f3bc1c2abb73173cd51ccabcc2520e922251d2a2bd66fa9ba91433e45df40756"} Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.779536 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3bc1c2abb73173cd51ccabcc2520e922251d2a2bd66fa9ba91433e45df40756" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.779565 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq-config-469c5" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.968146 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-chkks"] Oct 04 04:52:34 crc kubenswrapper[4992]: E1004 04:52:34.969619 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b8c7673-09e9-4b88-9970-5a3a01eac6a3" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.969640 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b8c7673-09e9-4b88-9970-5a3a01eac6a3" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: E1004 04:52:34.969699 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef0036a-de1e-40ee-99b3-5d15b2775955" containerName="ovn-config" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.969708 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef0036a-de1e-40ee-99b3-5d15b2775955" containerName="ovn-config" Oct 04 04:52:34 crc kubenswrapper[4992]: E1004 04:52:34.969722 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81a290cb-94bf-41ad-87bf-ccc8a713e03d" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.969737 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="81a290cb-94bf-41ad-87bf-ccc8a713e03d" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: E1004 04:52:34.969747 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd465ca0-c12b-4f3c-8de2-377ac6d81f89" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.969753 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd465ca0-c12b-4f3c-8de2-377ac6d81f89" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.969920 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ef0036a-de1e-40ee-99b3-5d15b2775955" containerName="ovn-config" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.969935 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="81a290cb-94bf-41ad-87bf-ccc8a713e03d" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.969946 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b8c7673-09e9-4b88-9970-5a3a01eac6a3" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.969959 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd465ca0-c12b-4f3c-8de2-377ac6d81f89" containerName="mariadb-account-create" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.970541 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-chkks" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.973646 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zjhxk" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.973771 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 04 04:52:34 crc kubenswrapper[4992]: I1004 04:52:34.975016 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-chkks"] Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.076328 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-db-sync-config-data\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.076429 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dzmk\" (UniqueName: \"kubernetes.io/projected/2ea89995-b28e-471b-8d9e-6f97d93fe073-kube-api-access-9dzmk\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.076591 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-combined-ca-bundle\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.076742 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-config-data\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.167182 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8m2tq-config-469c5"] Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.173839 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8m2tq-config-469c5"] Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.177839 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-db-sync-config-data\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.177875 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dzmk\" (UniqueName: \"kubernetes.io/projected/2ea89995-b28e-471b-8d9e-6f97d93fe073-kube-api-access-9dzmk\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.177913 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-combined-ca-bundle\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.177967 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-config-data\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.182388 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-config-data\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.183661 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-db-sync-config-data\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.184251 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-combined-ca-bundle\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.196157 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dzmk\" (UniqueName: \"kubernetes.io/projected/2ea89995-b28e-471b-8d9e-6f97d93fe073-kube-api-access-9dzmk\") pod \"glance-db-sync-chkks\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.285138 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-8m2tq-config-t2wqc"] Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.286744 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.290772 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.294496 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8m2tq-config-t2wqc"] Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.304226 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.353151 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-chkks" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.380666 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-scripts\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.380734 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run-ovn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.380817 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzjrn\" (UniqueName: \"kubernetes.io/projected/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-kube-api-access-nzjrn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.380870 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.380895 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-log-ovn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.380928 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-additional-scripts\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.482466 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.482523 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-log-ovn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.482558 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-additional-scripts\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.482649 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-scripts\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.482694 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run-ovn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.482782 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzjrn\" (UniqueName: \"kubernetes.io/projected/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-kube-api-access-nzjrn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.482857 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.483634 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-log-ovn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.484895 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-additional-scripts\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.485126 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run-ovn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.487724 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-scripts\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.503639 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzjrn\" (UniqueName: \"kubernetes.io/projected/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-kube-api-access-nzjrn\") pod \"ovn-controller-8m2tq-config-t2wqc\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.643974 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.778566 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.814794 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-8m2tq" Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.818651 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"d0bf075f44e18e09fd89db4b8c3cd5d3834d3005528de58a1286d0cd1e95d078"} Oct 04 04:52:35 crc kubenswrapper[4992]: I1004 04:52:35.869434 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-chkks"] Oct 04 04:52:35 crc kubenswrapper[4992]: W1004 04:52:35.879112 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ea89995_b28e_471b_8d9e_6f97d93fe073.slice/crio-bf7d6e2d39296c47478d6f1d5c2aacfa57ee7f732643cbe38fac97e23a66c053 WatchSource:0}: Error finding container bf7d6e2d39296c47478d6f1d5c2aacfa57ee7f732643cbe38fac97e23a66c053: Status 404 returned error can't find the container with id bf7d6e2d39296c47478d6f1d5c2aacfa57ee7f732643cbe38fac97e23a66c053 Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.054572 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.068031 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-hrlth"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.069034 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hrlth" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.153094 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hrlth"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.176068 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-lwcgl"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.181065 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lwcgl" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.204407 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lwcgl"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.207634 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l57fb\" (UniqueName: \"kubernetes.io/projected/291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44-kube-api-access-l57fb\") pod \"cinder-db-create-hrlth\" (UID: \"291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44\") " pod="openstack/cinder-db-create-hrlth" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.227646 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-8m2tq-config-t2wqc"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.309215 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l57fb\" (UniqueName: \"kubernetes.io/projected/291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44-kube-api-access-l57fb\") pod \"cinder-db-create-hrlth\" (UID: \"291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44\") " pod="openstack/cinder-db-create-hrlth" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.309289 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbpm5\" (UniqueName: \"kubernetes.io/projected/9c3904c3-0dc6-46b0-b152-688ce48cda67-kube-api-access-jbpm5\") pod \"barbican-db-create-lwcgl\" (UID: \"9c3904c3-0dc6-46b0-b152-688ce48cda67\") " pod="openstack/barbican-db-create-lwcgl" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.341332 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ef0036a-de1e-40ee-99b3-5d15b2775955" path="/var/lib/kubelet/pods/9ef0036a-de1e-40ee-99b3-5d15b2775955/volumes" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.381665 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l57fb\" (UniqueName: \"kubernetes.io/projected/291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44-kube-api-access-l57fb\") pod \"cinder-db-create-hrlth\" (UID: \"291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44\") " pod="openstack/cinder-db-create-hrlth" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.384662 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-2z2v2"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.385986 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2z2v2" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.399510 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2z2v2"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.410396 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbpm5\" (UniqueName: \"kubernetes.io/projected/9c3904c3-0dc6-46b0-b152-688ce48cda67-kube-api-access-jbpm5\") pod \"barbican-db-create-lwcgl\" (UID: \"9c3904c3-0dc6-46b0-b152-688ce48cda67\") " pod="openstack/barbican-db-create-lwcgl" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.437556 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hrlth" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.454843 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbpm5\" (UniqueName: \"kubernetes.io/projected/9c3904c3-0dc6-46b0-b152-688ce48cda67-kube-api-access-jbpm5\") pod \"barbican-db-create-lwcgl\" (UID: \"9c3904c3-0dc6-46b0-b152-688ce48cda67\") " pod="openstack/barbican-db-create-lwcgl" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.464803 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-22q9n"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.466023 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.468348 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.469724 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.471308 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.471503 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5nrbs" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.483277 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-22q9n"] Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.506876 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lwcgl" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.511838 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-config-data\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.512037 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxt2g\" (UniqueName: \"kubernetes.io/projected/99e52292-79e5-41bd-b360-40002e564a01-kube-api-access-pxt2g\") pod \"neutron-db-create-2z2v2\" (UID: \"99e52292-79e5-41bd-b360-40002e564a01\") " pod="openstack/neutron-db-create-2z2v2" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.512108 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-combined-ca-bundle\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.512162 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv7cl\" (UniqueName: \"kubernetes.io/projected/d5cba52e-0e38-4e46-98a6-4c08da5bd709-kube-api-access-dv7cl\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.614336 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxt2g\" (UniqueName: \"kubernetes.io/projected/99e52292-79e5-41bd-b360-40002e564a01-kube-api-access-pxt2g\") pod \"neutron-db-create-2z2v2\" (UID: \"99e52292-79e5-41bd-b360-40002e564a01\") " pod="openstack/neutron-db-create-2z2v2" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.614437 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-combined-ca-bundle\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.614502 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv7cl\" (UniqueName: \"kubernetes.io/projected/d5cba52e-0e38-4e46-98a6-4c08da5bd709-kube-api-access-dv7cl\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.614573 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-config-data\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.643977 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxt2g\" (UniqueName: \"kubernetes.io/projected/99e52292-79e5-41bd-b360-40002e564a01-kube-api-access-pxt2g\") pod \"neutron-db-create-2z2v2\" (UID: \"99e52292-79e5-41bd-b360-40002e564a01\") " pod="openstack/neutron-db-create-2z2v2" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.708422 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-combined-ca-bundle\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.708432 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-config-data\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.708677 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv7cl\" (UniqueName: \"kubernetes.io/projected/d5cba52e-0e38-4e46-98a6-4c08da5bd709-kube-api-access-dv7cl\") pod \"keystone-db-sync-22q9n\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.720010 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2z2v2" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.820883 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.827108 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq-config-t2wqc" event={"ID":"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed","Type":"ContainerStarted","Data":"d13a59a1df114a106065257bdc90f172626d7d877c5190bfb587393a78039c2f"} Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.827143 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq-config-t2wqc" event={"ID":"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed","Type":"ContainerStarted","Data":"c96b8cca1026e20c1c4a523b0c8cc0331346e7660a0270ad1b287381c9091d74"} Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.829731 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-chkks" event={"ID":"2ea89995-b28e-471b-8d9e-6f97d93fe073","Type":"ContainerStarted","Data":"bf7d6e2d39296c47478d6f1d5c2aacfa57ee7f732643cbe38fac97e23a66c053"} Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.960340 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-8m2tq-config-t2wqc" podStartSLOduration=1.960314315 podStartE2EDuration="1.960314315s" podCreationTimestamp="2025-10-04 04:52:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:52:36.846691296 +0000 UTC m=+1190.694366764" watchObservedRunningTime="2025-10-04 04:52:36.960314315 +0000 UTC m=+1190.807989783" Oct 04 04:52:36 crc kubenswrapper[4992]: I1004 04:52:36.967706 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-hrlth"] Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.058091 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lwcgl"] Oct 04 04:52:37 crc kubenswrapper[4992]: W1004 04:52:37.074000 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c3904c3_0dc6_46b0_b152_688ce48cda67.slice/crio-4487aee1c05733416d75bbe8e2a8f60d770ea320d95abfd3ded43e656b088120 WatchSource:0}: Error finding container 4487aee1c05733416d75bbe8e2a8f60d770ea320d95abfd3ded43e656b088120: Status 404 returned error can't find the container with id 4487aee1c05733416d75bbe8e2a8f60d770ea320d95abfd3ded43e656b088120 Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.612092 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2z2v2"] Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.631069 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-22q9n"] Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.860235 4992 generic.go:334] "Generic (PLEG): container finished" podID="9c3904c3-0dc6-46b0-b152-688ce48cda67" containerID="253c204ba74e7b725c88de437bc908f2df7f5c962d8532e5526e9dbc9693f034" exitCode=0 Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.860538 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lwcgl" event={"ID":"9c3904c3-0dc6-46b0-b152-688ce48cda67","Type":"ContainerDied","Data":"253c204ba74e7b725c88de437bc908f2df7f5c962d8532e5526e9dbc9693f034"} Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.860669 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lwcgl" event={"ID":"9c3904c3-0dc6-46b0-b152-688ce48cda67","Type":"ContainerStarted","Data":"4487aee1c05733416d75bbe8e2a8f60d770ea320d95abfd3ded43e656b088120"} Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.865958 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-22q9n" event={"ID":"d5cba52e-0e38-4e46-98a6-4c08da5bd709","Type":"ContainerStarted","Data":"1b06ae1051dba1467b77b419cfd706afabd3d2508e844b3bcb664ce3d3edceab"} Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.925256 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"6e7ceae3d24481394dea8b438442b938ffd19546d24666db360e75c3d497d4cd"} Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.925313 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"6fde98b34927098e0929e161f1ee17612e3c13aeeb912c584826b280d0658040"} Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.937502 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2z2v2" event={"ID":"99e52292-79e5-41bd-b360-40002e564a01","Type":"ContainerStarted","Data":"5aaac4e3fb4aff0c546f997c2c710368bb0c00da4e4e25e485a869f69604a676"} Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.952565 4992 generic.go:334] "Generic (PLEG): container finished" podID="291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44" containerID="4e88a53ee59cfa6d83eee2666d49888fe8b2437146c38b633dffcaf41a43e83f" exitCode=0 Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.953008 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hrlth" event={"ID":"291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44","Type":"ContainerDied","Data":"4e88a53ee59cfa6d83eee2666d49888fe8b2437146c38b633dffcaf41a43e83f"} Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.953047 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hrlth" event={"ID":"291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44","Type":"ContainerStarted","Data":"e00b0614f087f65f310957f89299fb56e0215ada3ab9b0cdd9f9f33aff4595cb"} Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.963724 4992 generic.go:334] "Generic (PLEG): container finished" podID="0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" containerID="d13a59a1df114a106065257bdc90f172626d7d877c5190bfb587393a78039c2f" exitCode=0 Oct 04 04:52:37 crc kubenswrapper[4992]: I1004 04:52:37.963775 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq-config-t2wqc" event={"ID":"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed","Type":"ContainerDied","Data":"d13a59a1df114a106065257bdc90f172626d7d877c5190bfb587393a78039c2f"} Oct 04 04:52:38 crc kubenswrapper[4992]: I1004 04:52:38.976863 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"7cfb9b84ff0fa7fe59aec8b2343f1323fc8bd126fc456cea6cc6ebbcf43acaaa"} Oct 04 04:52:38 crc kubenswrapper[4992]: I1004 04:52:38.977060 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"828a2d2d6138edcc3257c7e477b3008a86105cc778476623177bd8177743cc2a"} Oct 04 04:52:38 crc kubenswrapper[4992]: I1004 04:52:38.978486 4992 generic.go:334] "Generic (PLEG): container finished" podID="99e52292-79e5-41bd-b360-40002e564a01" containerID="451f9b9f959f53d1ddba9019564ba6b46acf04f31cc977d4ab8d130873fc7056" exitCode=0 Oct 04 04:52:38 crc kubenswrapper[4992]: I1004 04:52:38.978562 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2z2v2" event={"ID":"99e52292-79e5-41bd-b360-40002e564a01","Type":"ContainerDied","Data":"451f9b9f959f53d1ddba9019564ba6b46acf04f31cc977d4ab8d130873fc7056"} Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.424295 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lwcgl" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.442423 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hrlth" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.452186 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.466259 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-additional-scripts\") pod \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.466693 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbpm5\" (UniqueName: \"kubernetes.io/projected/9c3904c3-0dc6-46b0-b152-688ce48cda67-kube-api-access-jbpm5\") pod \"9c3904c3-0dc6-46b0-b152-688ce48cda67\" (UID: \"9c3904c3-0dc6-46b0-b152-688ce48cda67\") " Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.466858 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-log-ovn\") pod \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.466929 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" (UID: "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.466954 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run-ovn\") pod \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.467169 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l57fb\" (UniqueName: \"kubernetes.io/projected/291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44-kube-api-access-l57fb\") pod \"291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44\" (UID: \"291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44\") " Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.467306 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzjrn\" (UniqueName: \"kubernetes.io/projected/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-kube-api-access-nzjrn\") pod \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.471725 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run\") pod \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.467050 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" (UID: "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.467191 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" (UID: "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.471877 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run" (OuterVolumeSpecName: "var-run") pod "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" (UID: "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.472149 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-scripts\") pod \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\" (UID: \"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed\") " Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.472823 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-scripts" (OuterVolumeSpecName: "scripts") pod "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" (UID: "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.473114 4992 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.473314 4992 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.473432 4992 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.473570 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.473699 4992 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.474922 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c3904c3-0dc6-46b0-b152-688ce48cda67-kube-api-access-jbpm5" (OuterVolumeSpecName: "kube-api-access-jbpm5") pod "9c3904c3-0dc6-46b0-b152-688ce48cda67" (UID: "9c3904c3-0dc6-46b0-b152-688ce48cda67"). InnerVolumeSpecName "kube-api-access-jbpm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.476243 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-kube-api-access-nzjrn" (OuterVolumeSpecName: "kube-api-access-nzjrn") pod "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" (UID: "0d6dc325-6c36-4cb5-af09-fe3cd389b5ed"). InnerVolumeSpecName "kube-api-access-nzjrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.479297 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44-kube-api-access-l57fb" (OuterVolumeSpecName: "kube-api-access-l57fb") pod "291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44" (UID: "291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44"). InnerVolumeSpecName "kube-api-access-l57fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.574768 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l57fb\" (UniqueName: \"kubernetes.io/projected/291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44-kube-api-access-l57fb\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.574797 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzjrn\" (UniqueName: \"kubernetes.io/projected/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed-kube-api-access-nzjrn\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.574806 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbpm5\" (UniqueName: \"kubernetes.io/projected/9c3904c3-0dc6-46b0-b152-688ce48cda67-kube-api-access-jbpm5\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.990235 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-hrlth" event={"ID":"291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44","Type":"ContainerDied","Data":"e00b0614f087f65f310957f89299fb56e0215ada3ab9b0cdd9f9f33aff4595cb"} Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.990282 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e00b0614f087f65f310957f89299fb56e0215ada3ab9b0cdd9f9f33aff4595cb" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.990256 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-hrlth" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.992126 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-8m2tq-config-t2wqc" event={"ID":"0d6dc325-6c36-4cb5-af09-fe3cd389b5ed","Type":"ContainerDied","Data":"c96b8cca1026e20c1c4a523b0c8cc0331346e7660a0270ad1b287381c9091d74"} Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.992148 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c96b8cca1026e20c1c4a523b0c8cc0331346e7660a0270ad1b287381c9091d74" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.992180 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-8m2tq-config-t2wqc" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.993927 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lwcgl" event={"ID":"9c3904c3-0dc6-46b0-b152-688ce48cda67","Type":"ContainerDied","Data":"4487aee1c05733416d75bbe8e2a8f60d770ea320d95abfd3ded43e656b088120"} Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.993959 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4487aee1c05733416d75bbe8e2a8f60d770ea320d95abfd3ded43e656b088120" Oct 04 04:52:39 crc kubenswrapper[4992]: I1004 04:52:39.994061 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lwcgl" Oct 04 04:52:40 crc kubenswrapper[4992]: I1004 04:52:40.531139 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-8m2tq-config-t2wqc"] Oct 04 04:52:40 crc kubenswrapper[4992]: I1004 04:52:40.538372 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-8m2tq-config-t2wqc"] Oct 04 04:52:40 crc kubenswrapper[4992]: I1004 04:52:40.980397 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2z2v2" Oct 04 04:52:41 crc kubenswrapper[4992]: I1004 04:52:41.002381 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxt2g\" (UniqueName: \"kubernetes.io/projected/99e52292-79e5-41bd-b360-40002e564a01-kube-api-access-pxt2g\") pod \"99e52292-79e5-41bd-b360-40002e564a01\" (UID: \"99e52292-79e5-41bd-b360-40002e564a01\") " Oct 04 04:52:41 crc kubenswrapper[4992]: I1004 04:52:41.021037 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e52292-79e5-41bd-b360-40002e564a01-kube-api-access-pxt2g" (OuterVolumeSpecName: "kube-api-access-pxt2g") pod "99e52292-79e5-41bd-b360-40002e564a01" (UID: "99e52292-79e5-41bd-b360-40002e564a01"). InnerVolumeSpecName "kube-api-access-pxt2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:41 crc kubenswrapper[4992]: I1004 04:52:41.028833 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2z2v2" event={"ID":"99e52292-79e5-41bd-b360-40002e564a01","Type":"ContainerDied","Data":"5aaac4e3fb4aff0c546f997c2c710368bb0c00da4e4e25e485a869f69604a676"} Oct 04 04:52:41 crc kubenswrapper[4992]: I1004 04:52:41.028884 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5aaac4e3fb4aff0c546f997c2c710368bb0c00da4e4e25e485a869f69604a676" Oct 04 04:52:41 crc kubenswrapper[4992]: I1004 04:52:41.028979 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2z2v2" Oct 04 04:52:41 crc kubenswrapper[4992]: I1004 04:52:41.104496 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxt2g\" (UniqueName: \"kubernetes.io/projected/99e52292-79e5-41bd-b360-40002e564a01-kube-api-access-pxt2g\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:42 crc kubenswrapper[4992]: I1004 04:52:42.330512 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" path="/var/lib/kubelet/pods/0d6dc325-6c36-4cb5-af09-fe3cd389b5ed/volumes" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.216288 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-204f-account-create-wpsgm"] Oct 04 04:52:46 crc kubenswrapper[4992]: E1004 04:52:46.216870 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3904c3-0dc6-46b0-b152-688ce48cda67" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.216884 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3904c3-0dc6-46b0-b152-688ce48cda67" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: E1004 04:52:46.216892 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.216899 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: E1004 04:52:46.216921 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e52292-79e5-41bd-b360-40002e564a01" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.216927 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e52292-79e5-41bd-b360-40002e564a01" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: E1004 04:52:46.216943 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" containerName="ovn-config" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.216949 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" containerName="ovn-config" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.217105 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3904c3-0dc6-46b0-b152-688ce48cda67" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.217149 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d6dc325-6c36-4cb5-af09-fe3cd389b5ed" containerName="ovn-config" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.217174 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e52292-79e5-41bd-b360-40002e564a01" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.217230 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44" containerName="mariadb-database-create" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.220512 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-204f-account-create-wpsgm" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.222868 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.234676 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-204f-account-create-wpsgm"] Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.293912 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf5dw\" (UniqueName: \"kubernetes.io/projected/718cab8d-53e1-4d02-940a-4b2882070fa6-kube-api-access-mf5dw\") pod \"cinder-204f-account-create-wpsgm\" (UID: \"718cab8d-53e1-4d02-940a-4b2882070fa6\") " pod="openstack/cinder-204f-account-create-wpsgm" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.306094 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-095f-account-create-vdc9q"] Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.307437 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-095f-account-create-vdc9q" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.313905 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-095f-account-create-vdc9q"] Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.338904 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.395951 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r644s\" (UniqueName: \"kubernetes.io/projected/7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd-kube-api-access-r644s\") pod \"barbican-095f-account-create-vdc9q\" (UID: \"7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd\") " pod="openstack/barbican-095f-account-create-vdc9q" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.396035 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf5dw\" (UniqueName: \"kubernetes.io/projected/718cab8d-53e1-4d02-940a-4b2882070fa6-kube-api-access-mf5dw\") pod \"cinder-204f-account-create-wpsgm\" (UID: \"718cab8d-53e1-4d02-940a-4b2882070fa6\") " pod="openstack/cinder-204f-account-create-wpsgm" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.418483 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf5dw\" (UniqueName: \"kubernetes.io/projected/718cab8d-53e1-4d02-940a-4b2882070fa6-kube-api-access-mf5dw\") pod \"cinder-204f-account-create-wpsgm\" (UID: \"718cab8d-53e1-4d02-940a-4b2882070fa6\") " pod="openstack/cinder-204f-account-create-wpsgm" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.497886 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r644s\" (UniqueName: \"kubernetes.io/projected/7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd-kube-api-access-r644s\") pod \"barbican-095f-account-create-vdc9q\" (UID: \"7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd\") " pod="openstack/barbican-095f-account-create-vdc9q" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.513556 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r644s\" (UniqueName: \"kubernetes.io/projected/7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd-kube-api-access-r644s\") pod \"barbican-095f-account-create-vdc9q\" (UID: \"7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd\") " pod="openstack/barbican-095f-account-create-vdc9q" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.543571 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-204f-account-create-wpsgm" Oct 04 04:52:46 crc kubenswrapper[4992]: I1004 04:52:46.659404 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-095f-account-create-vdc9q" Oct 04 04:52:51 crc kubenswrapper[4992]: I1004 04:52:51.553321 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-204f-account-create-wpsgm"] Oct 04 04:52:51 crc kubenswrapper[4992]: I1004 04:52:51.558856 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-095f-account-create-vdc9q"] Oct 04 04:52:51 crc kubenswrapper[4992]: W1004 04:52:51.567182 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b4d3a94_3c31_49a6_9f0b_4b6f7cd227dd.slice/crio-64dcf6f1b116bbebfe30d5fb9f84921784129ff48d03f26134d48a32045a3034 WatchSource:0}: Error finding container 64dcf6f1b116bbebfe30d5fb9f84921784129ff48d03f26134d48a32045a3034: Status 404 returned error can't find the container with id 64dcf6f1b116bbebfe30d5fb9f84921784129ff48d03f26134d48a32045a3034 Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.129877 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-22q9n" event={"ID":"d5cba52e-0e38-4e46-98a6-4c08da5bd709","Type":"ContainerStarted","Data":"c540711e4eb3ba9d17845bc1780e6a24be38964532fb2530c93b2cc32f1cbbbe"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.134236 4992 generic.go:334] "Generic (PLEG): container finished" podID="7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd" containerID="c68804cb71e80e3632d31b028a7a4b55259bcd8520ba5ba9d5efe980650a0d80" exitCode=0 Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.134464 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-095f-account-create-vdc9q" event={"ID":"7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd","Type":"ContainerDied","Data":"c68804cb71e80e3632d31b028a7a4b55259bcd8520ba5ba9d5efe980650a0d80"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.134499 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-095f-account-create-vdc9q" event={"ID":"7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd","Type":"ContainerStarted","Data":"64dcf6f1b116bbebfe30d5fb9f84921784129ff48d03f26134d48a32045a3034"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.140289 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"d9b673408ac30f4884e863134b1b69f014f3ba608ab5af9a9f7c8f0f64cbaa19"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.140336 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"96273f9d5f8276ec677c97f5d26c928c955f0c87104b7ba9f91ff57001d5e877"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.140347 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"e74a2dc873ef278cf4ecd5f7a3db37afebb09b324d9a264589858e499da02c29"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.140373 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"51d91757648d3c4fac2ebe85b6bcfd1543c97f4f43f49b35d85005874fe34b44"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.141987 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-chkks" event={"ID":"2ea89995-b28e-471b-8d9e-6f97d93fe073","Type":"ContainerStarted","Data":"33de08d820fa120fa3e4c06b5ec746bda80c31ac74728f3c9c351b9bf084588c"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.147241 4992 generic.go:334] "Generic (PLEG): container finished" podID="718cab8d-53e1-4d02-940a-4b2882070fa6" containerID="a42fa51e630aa00f894fdc19ce2157d48d9ee5d7f77e4e4c756543c343c66f00" exitCode=0 Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.147306 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-204f-account-create-wpsgm" event={"ID":"718cab8d-53e1-4d02-940a-4b2882070fa6","Type":"ContainerDied","Data":"a42fa51e630aa00f894fdc19ce2157d48d9ee5d7f77e4e4c756543c343c66f00"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.147336 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-204f-account-create-wpsgm" event={"ID":"718cab8d-53e1-4d02-940a-4b2882070fa6","Type":"ContainerStarted","Data":"b07bb1a1cac70863fdfcb3c671dead392b6b9ba3b8c29d0832aa67d0d806e7cd"} Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.158864 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-22q9n" podStartSLOduration=2.696706173 podStartE2EDuration="16.158842819s" podCreationTimestamp="2025-10-04 04:52:36 +0000 UTC" firstStartedPulling="2025-10-04 04:52:37.64489384 +0000 UTC m=+1191.492569308" lastFinishedPulling="2025-10-04 04:52:51.107030476 +0000 UTC m=+1204.954705954" observedRunningTime="2025-10-04 04:52:52.149394037 +0000 UTC m=+1205.997069505" watchObservedRunningTime="2025-10-04 04:52:52.158842819 +0000 UTC m=+1206.006518287" Oct 04 04:52:52 crc kubenswrapper[4992]: I1004 04:52:52.170097 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-chkks" podStartSLOduration=2.933604072 podStartE2EDuration="18.170077139s" podCreationTimestamp="2025-10-04 04:52:34 +0000 UTC" firstStartedPulling="2025-10-04 04:52:35.881647595 +0000 UTC m=+1189.729323063" lastFinishedPulling="2025-10-04 04:52:51.118120662 +0000 UTC m=+1204.965796130" observedRunningTime="2025-10-04 04:52:52.168881667 +0000 UTC m=+1206.016557135" watchObservedRunningTime="2025-10-04 04:52:52.170077139 +0000 UTC m=+1206.017752607" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.064830 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-095f-account-create-vdc9q" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.071798 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-204f-account-create-wpsgm" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.145086 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf5dw\" (UniqueName: \"kubernetes.io/projected/718cab8d-53e1-4d02-940a-4b2882070fa6-kube-api-access-mf5dw\") pod \"718cab8d-53e1-4d02-940a-4b2882070fa6\" (UID: \"718cab8d-53e1-4d02-940a-4b2882070fa6\") " Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.145147 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r644s\" (UniqueName: \"kubernetes.io/projected/7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd-kube-api-access-r644s\") pod \"7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd\" (UID: \"7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd\") " Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.151561 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd-kube-api-access-r644s" (OuterVolumeSpecName: "kube-api-access-r644s") pod "7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd" (UID: "7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd"). InnerVolumeSpecName "kube-api-access-r644s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.151656 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/718cab8d-53e1-4d02-940a-4b2882070fa6-kube-api-access-mf5dw" (OuterVolumeSpecName: "kube-api-access-mf5dw") pod "718cab8d-53e1-4d02-940a-4b2882070fa6" (UID: "718cab8d-53e1-4d02-940a-4b2882070fa6"). InnerVolumeSpecName "kube-api-access-mf5dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.177387 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-204f-account-create-wpsgm" event={"ID":"718cab8d-53e1-4d02-940a-4b2882070fa6","Type":"ContainerDied","Data":"b07bb1a1cac70863fdfcb3c671dead392b6b9ba3b8c29d0832aa67d0d806e7cd"} Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.177435 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b07bb1a1cac70863fdfcb3c671dead392b6b9ba3b8c29d0832aa67d0d806e7cd" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.177508 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-204f-account-create-wpsgm" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.179336 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-095f-account-create-vdc9q" event={"ID":"7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd","Type":"ContainerDied","Data":"64dcf6f1b116bbebfe30d5fb9f84921784129ff48d03f26134d48a32045a3034"} Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.179375 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64dcf6f1b116bbebfe30d5fb9f84921784129ff48d03f26134d48a32045a3034" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.179421 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-095f-account-create-vdc9q" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.246843 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf5dw\" (UniqueName: \"kubernetes.io/projected/718cab8d-53e1-4d02-940a-4b2882070fa6-kube-api-access-mf5dw\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:55 crc kubenswrapper[4992]: I1004 04:52:55.246873 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r644s\" (UniqueName: \"kubernetes.io/projected/7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd-kube-api-access-r644s\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.203093 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"c296eeaf0612ce7b9bf1e077a6c03818b40dc0908193db3955411d418831f7fb"} Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.203403 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"5c21f5e5fb1f59a641c2613446bdafbed28e1fc573110ad50f268014e9c4d98b"} Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.203419 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"cabf0354c4384cddb4096b538e113a41e8f9ae61401c80feafa486dfbc7ef22c"} Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.203429 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"572591a772d87c1acdb3745c3bff404df7a9ad14318dd8d60e120ecacec65db7"} Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.203438 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"b332b247d18fb6d9ccd401dbd4e683d418f6a444caa3ff4d43bece71ccacf72b"} Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.547594 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-cc11-account-create-svpcv"] Oct 04 04:52:56 crc kubenswrapper[4992]: E1004 04:52:56.548241 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="718cab8d-53e1-4d02-940a-4b2882070fa6" containerName="mariadb-account-create" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.548267 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="718cab8d-53e1-4d02-940a-4b2882070fa6" containerName="mariadb-account-create" Oct 04 04:52:56 crc kubenswrapper[4992]: E1004 04:52:56.548288 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd" containerName="mariadb-account-create" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.548296 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd" containerName="mariadb-account-create" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.548559 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd" containerName="mariadb-account-create" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.548600 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="718cab8d-53e1-4d02-940a-4b2882070fa6" containerName="mariadb-account-create" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.549232 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc11-account-create-svpcv" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.551792 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.555010 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cc11-account-create-svpcv"] Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.569633 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44tck\" (UniqueName: \"kubernetes.io/projected/2b1cd82a-9ceb-4a24-937b-b4917eaa473a-kube-api-access-44tck\") pod \"neutron-cc11-account-create-svpcv\" (UID: \"2b1cd82a-9ceb-4a24-937b-b4917eaa473a\") " pod="openstack/neutron-cc11-account-create-svpcv" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.671229 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44tck\" (UniqueName: \"kubernetes.io/projected/2b1cd82a-9ceb-4a24-937b-b4917eaa473a-kube-api-access-44tck\") pod \"neutron-cc11-account-create-svpcv\" (UID: \"2b1cd82a-9ceb-4a24-937b-b4917eaa473a\") " pod="openstack/neutron-cc11-account-create-svpcv" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.688158 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44tck\" (UniqueName: \"kubernetes.io/projected/2b1cd82a-9ceb-4a24-937b-b4917eaa473a-kube-api-access-44tck\") pod \"neutron-cc11-account-create-svpcv\" (UID: \"2b1cd82a-9ceb-4a24-937b-b4917eaa473a\") " pod="openstack/neutron-cc11-account-create-svpcv" Oct 04 04:52:56 crc kubenswrapper[4992]: I1004 04:52:56.865217 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc11-account-create-svpcv" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.218243 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"a419339a04ce974ace67bb104dd4b2759c0f7d4b2803edb0698098e7d04123e8"} Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.218549 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"fa875baa-782c-469c-816f-7727c4885d28","Type":"ContainerStarted","Data":"4cac6e0d6547031b26ae5f49239617acd2bb417dd0d8f39765d87b39180250c7"} Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.253627 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.514848234 podStartE2EDuration="56.253607095s" podCreationTimestamp="2025-10-04 04:52:01 +0000 UTC" firstStartedPulling="2025-10-04 04:52:35.301414671 +0000 UTC m=+1189.149090139" lastFinishedPulling="2025-10-04 04:52:55.040173522 +0000 UTC m=+1208.887849000" observedRunningTime="2025-10-04 04:52:57.246941827 +0000 UTC m=+1211.094617305" watchObservedRunningTime="2025-10-04 04:52:57.253607095 +0000 UTC m=+1211.101282563" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.298785 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cc11-account-create-svpcv"] Oct 04 04:52:57 crc kubenswrapper[4992]: W1004 04:52:57.300615 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b1cd82a_9ceb_4a24_937b_b4917eaa473a.slice/crio-816a3bc2d2e47769d6f1c8976d77c452a2873d62fc253f7ebe72b962198a4d6e WatchSource:0}: Error finding container 816a3bc2d2e47769d6f1c8976d77c452a2873d62fc253f7ebe72b962198a4d6e: Status 404 returned error can't find the container with id 816a3bc2d2e47769d6f1c8976d77c452a2873d62fc253f7ebe72b962198a4d6e Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.520626 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x4zbx"] Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.522686 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.524580 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.538822 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x4zbx"] Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.585816 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.585870 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.585961 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.585994 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-config\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.586031 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gqxg\" (UniqueName: \"kubernetes.io/projected/84d13164-e16f-4a6a-b8a0-5bcba00d4649-kube-api-access-4gqxg\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.586154 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.687405 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.687467 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-config\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.687528 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gqxg\" (UniqueName: \"kubernetes.io/projected/84d13164-e16f-4a6a-b8a0-5bcba00d4649-kube-api-access-4gqxg\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.687605 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.687636 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.687667 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.689120 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.689267 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.689487 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.689496 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.689686 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-config\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.710298 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gqxg\" (UniqueName: \"kubernetes.io/projected/84d13164-e16f-4a6a-b8a0-5bcba00d4649-kube-api-access-4gqxg\") pod \"dnsmasq-dns-77585f5f8c-x4zbx\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:57 crc kubenswrapper[4992]: I1004 04:52:57.840889 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:52:58 crc kubenswrapper[4992]: I1004 04:52:58.228517 4992 generic.go:334] "Generic (PLEG): container finished" podID="d5cba52e-0e38-4e46-98a6-4c08da5bd709" containerID="c540711e4eb3ba9d17845bc1780e6a24be38964532fb2530c93b2cc32f1cbbbe" exitCode=0 Oct 04 04:52:58 crc kubenswrapper[4992]: I1004 04:52:58.228556 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-22q9n" event={"ID":"d5cba52e-0e38-4e46-98a6-4c08da5bd709","Type":"ContainerDied","Data":"c540711e4eb3ba9d17845bc1780e6a24be38964532fb2530c93b2cc32f1cbbbe"} Oct 04 04:52:58 crc kubenswrapper[4992]: I1004 04:52:58.230929 4992 generic.go:334] "Generic (PLEG): container finished" podID="2b1cd82a-9ceb-4a24-937b-b4917eaa473a" containerID="e067dd9c744fda9f058e838f00a3439d2616d1a94f18522d82fdf28c9eae8e53" exitCode=0 Oct 04 04:52:58 crc kubenswrapper[4992]: I1004 04:52:58.231055 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc11-account-create-svpcv" event={"ID":"2b1cd82a-9ceb-4a24-937b-b4917eaa473a","Type":"ContainerDied","Data":"e067dd9c744fda9f058e838f00a3439d2616d1a94f18522d82fdf28c9eae8e53"} Oct 04 04:52:58 crc kubenswrapper[4992]: I1004 04:52:58.231097 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc11-account-create-svpcv" event={"ID":"2b1cd82a-9ceb-4a24-937b-b4917eaa473a","Type":"ContainerStarted","Data":"816a3bc2d2e47769d6f1c8976d77c452a2873d62fc253f7ebe72b962198a4d6e"} Oct 04 04:52:58 crc kubenswrapper[4992]: I1004 04:52:58.275910 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x4zbx"] Oct 04 04:52:58 crc kubenswrapper[4992]: W1004 04:52:58.284565 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84d13164_e16f_4a6a_b8a0_5bcba00d4649.slice/crio-05b08aa266af718f2121b56d26f31c2304129d43b710ce460d40ae1b8c84768a WatchSource:0}: Error finding container 05b08aa266af718f2121b56d26f31c2304129d43b710ce460d40ae1b8c84768a: Status 404 returned error can't find the container with id 05b08aa266af718f2121b56d26f31c2304129d43b710ce460d40ae1b8c84768a Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.240149 4992 generic.go:334] "Generic (PLEG): container finished" podID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" containerID="00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602" exitCode=0 Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.240451 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" event={"ID":"84d13164-e16f-4a6a-b8a0-5bcba00d4649","Type":"ContainerDied","Data":"00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602"} Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.240889 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" event={"ID":"84d13164-e16f-4a6a-b8a0-5bcba00d4649","Type":"ContainerStarted","Data":"05b08aa266af718f2121b56d26f31c2304129d43b710ce460d40ae1b8c84768a"} Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.589988 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc11-account-create-svpcv" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.595588 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-22q9n" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.720894 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44tck\" (UniqueName: \"kubernetes.io/projected/2b1cd82a-9ceb-4a24-937b-b4917eaa473a-kube-api-access-44tck\") pod \"2b1cd82a-9ceb-4a24-937b-b4917eaa473a\" (UID: \"2b1cd82a-9ceb-4a24-937b-b4917eaa473a\") " Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.721065 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-combined-ca-bundle\") pod \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.721114 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-config-data\") pod \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.721153 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv7cl\" (UniqueName: \"kubernetes.io/projected/d5cba52e-0e38-4e46-98a6-4c08da5bd709-kube-api-access-dv7cl\") pod \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\" (UID: \"d5cba52e-0e38-4e46-98a6-4c08da5bd709\") " Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.726610 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b1cd82a-9ceb-4a24-937b-b4917eaa473a-kube-api-access-44tck" (OuterVolumeSpecName: "kube-api-access-44tck") pod "2b1cd82a-9ceb-4a24-937b-b4917eaa473a" (UID: "2b1cd82a-9ceb-4a24-937b-b4917eaa473a"). InnerVolumeSpecName "kube-api-access-44tck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.726837 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5cba52e-0e38-4e46-98a6-4c08da5bd709-kube-api-access-dv7cl" (OuterVolumeSpecName: "kube-api-access-dv7cl") pod "d5cba52e-0e38-4e46-98a6-4c08da5bd709" (UID: "d5cba52e-0e38-4e46-98a6-4c08da5bd709"). InnerVolumeSpecName "kube-api-access-dv7cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.751156 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5cba52e-0e38-4e46-98a6-4c08da5bd709" (UID: "d5cba52e-0e38-4e46-98a6-4c08da5bd709"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.762992 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-config-data" (OuterVolumeSpecName: "config-data") pod "d5cba52e-0e38-4e46-98a6-4c08da5bd709" (UID: "d5cba52e-0e38-4e46-98a6-4c08da5bd709"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.823462 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.823498 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5cba52e-0e38-4e46-98a6-4c08da5bd709-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.823508 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv7cl\" (UniqueName: \"kubernetes.io/projected/d5cba52e-0e38-4e46-98a6-4c08da5bd709-kube-api-access-dv7cl\") on node \"crc\" DevicePath \"\"" Oct 04 04:52:59 crc kubenswrapper[4992]: I1004 04:52:59.823519 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44tck\" (UniqueName: \"kubernetes.io/projected/2b1cd82a-9ceb-4a24-937b-b4917eaa473a-kube-api-access-44tck\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.250319 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" event={"ID":"84d13164-e16f-4a6a-b8a0-5bcba00d4649","Type":"ContainerStarted","Data":"1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63"} Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.251316 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.252888 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-22q9n" event={"ID":"d5cba52e-0e38-4e46-98a6-4c08da5bd709","Type":"ContainerDied","Data":"1b06ae1051dba1467b77b419cfd706afabd3d2508e844b3bcb664ce3d3edceab"} Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.252923 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b06ae1051dba1467b77b419cfd706afabd3d2508e844b3bcb664ce3d3edceab" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.252965 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-22q9n" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.264319 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc11-account-create-svpcv" event={"ID":"2b1cd82a-9ceb-4a24-937b-b4917eaa473a","Type":"ContainerDied","Data":"816a3bc2d2e47769d6f1c8976d77c452a2873d62fc253f7ebe72b962198a4d6e"} Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.264648 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="816a3bc2d2e47769d6f1c8976d77c452a2873d62fc253f7ebe72b962198a4d6e" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.264800 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc11-account-create-svpcv" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.284527 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" podStartSLOduration=3.284507505 podStartE2EDuration="3.284507505s" podCreationTimestamp="2025-10-04 04:52:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:00.282011689 +0000 UTC m=+1214.129687167" watchObservedRunningTime="2025-10-04 04:53:00.284507505 +0000 UTC m=+1214.132182973" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.475828 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x4zbx"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.521527 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-hh6m8"] Oct 04 04:53:00 crc kubenswrapper[4992]: E1004 04:53:00.522318 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5cba52e-0e38-4e46-98a6-4c08da5bd709" containerName="keystone-db-sync" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.522372 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5cba52e-0e38-4e46-98a6-4c08da5bd709" containerName="keystone-db-sync" Oct 04 04:53:00 crc kubenswrapper[4992]: E1004 04:53:00.522385 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1cd82a-9ceb-4a24-937b-b4917eaa473a" containerName="mariadb-account-create" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.522394 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1cd82a-9ceb-4a24-937b-b4917eaa473a" containerName="mariadb-account-create" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.522595 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b1cd82a-9ceb-4a24-937b-b4917eaa473a" containerName="mariadb-account-create" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.522612 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5cba52e-0e38-4e46-98a6-4c08da5bd709" containerName="keystone-db-sync" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.523748 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.531352 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-hh6m8"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.539034 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-config\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.539099 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.539144 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb7kl\" (UniqueName: \"kubernetes.io/projected/aaa6ec0a-ce61-49e1-9e4b-78639257598f-kube-api-access-zb7kl\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.539454 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.539540 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-svc\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.539697 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.540288 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4vnm9"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.541536 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.543217 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5nrbs" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.544204 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.545074 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.545727 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.586790 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4vnm9"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.640935 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-fernet-keys\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.640995 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641019 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddmjt\" (UniqueName: \"kubernetes.io/projected/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-kube-api-access-ddmjt\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641057 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb7kl\" (UniqueName: \"kubernetes.io/projected/aaa6ec0a-ce61-49e1-9e4b-78639257598f-kube-api-access-zb7kl\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641085 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-scripts\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641131 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641148 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-svc\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641169 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-config-data\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641189 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641222 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-credential-keys\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641242 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-config\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.641261 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-combined-ca-bundle\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.642115 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.642872 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.643456 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-svc\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.643767 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.643998 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-config\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.681068 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-zcz9v"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.682718 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.687770 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9h6dr" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.688096 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.688215 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.688307 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-9c7f9b5b9-6878s"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.713785 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb7kl\" (UniqueName: \"kubernetes.io/projected/aaa6ec0a-ce61-49e1-9e4b-78639257598f-kube-api-access-zb7kl\") pod \"dnsmasq-dns-55fff446b9-hh6m8\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.742966 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-scripts\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743046 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-config-data\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743080 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-scripts\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743117 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-config-data\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743139 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-db-sync-config-data\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743158 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-combined-ca-bundle\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743191 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-credential-keys\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743217 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-combined-ca-bundle\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743242 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-etc-machine-id\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743258 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpwvw\" (UniqueName: \"kubernetes.io/projected/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-kube-api-access-wpwvw\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743280 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-fernet-keys\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.743328 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddmjt\" (UniqueName: \"kubernetes.io/projected/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-kube-api-access-ddmjt\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.747060 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-zcz9v"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.747205 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.772461 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-fernet-keys\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.777040 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-config-data\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.781261 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-t2h9s" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.784916 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.785118 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.785214 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.785858 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-credential-keys\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.791179 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-combined-ca-bundle\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.804732 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-scripts\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.807241 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddmjt\" (UniqueName: \"kubernetes.io/projected/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-kube-api-access-ddmjt\") pod \"keystone-bootstrap-4vnm9\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.829818 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9c7f9b5b9-6878s"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.845192 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-db-sync-config-data\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.845236 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-combined-ca-bundle\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.845299 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-etc-machine-id\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.845317 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpwvw\" (UniqueName: \"kubernetes.io/projected/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-kube-api-access-wpwvw\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.845405 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-config-data\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.845424 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-scripts\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.856838 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.857906 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-etc-machine-id\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.861697 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.878982 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-db-sync-config-data\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.879386 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.882143 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-config-data\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.882425 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.888482 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-combined-ca-bundle\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.892894 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-scripts\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.904264 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.904674 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.920890 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpwvw\" (UniqueName: \"kubernetes.io/projected/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-kube-api-access-wpwvw\") pod \"cinder-db-sync-zcz9v\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.930536 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.949050 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-config-data\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.949093 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76056377-cff7-4d4c-b463-8c25a7b0e48e-logs\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.949136 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-scripts\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.949152 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfhwg\" (UniqueName: \"kubernetes.io/projected/76056377-cff7-4d4c-b463-8c25a7b0e48e-kube-api-access-qfhwg\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:00 crc kubenswrapper[4992]: I1004 04:53:00.949187 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76056377-cff7-4d4c-b463-8c25a7b0e48e-horizon-secret-key\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.044577 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-srrcj"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.046011 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.057947 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fvv86" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058312 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058569 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-config-data\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058606 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76056377-cff7-4d4c-b463-8c25a7b0e48e-logs\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058652 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058690 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-scripts\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058709 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfhwg\" (UniqueName: \"kubernetes.io/projected/76056377-cff7-4d4c-b463-8c25a7b0e48e-kube-api-access-qfhwg\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058741 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-scripts\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058755 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058783 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-config-data\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058800 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76056377-cff7-4d4c-b463-8c25a7b0e48e-horizon-secret-key\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058828 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-log-httpd\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058852 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx6tl\" (UniqueName: \"kubernetes.io/projected/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-kube-api-access-fx6tl\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.058880 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-run-httpd\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.059092 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76056377-cff7-4d4c-b463-8c25a7b0e48e-logs\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.059454 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-scripts\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.061630 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-config-data\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.076828 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76056377-cff7-4d4c-b463-8c25a7b0e48e-horizon-secret-key\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.084952 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-5578x"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.086326 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.093222 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.093488 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-cb2c8" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.093631 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.112718 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5578x"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.118987 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfhwg\" (UniqueName: \"kubernetes.io/projected/76056377-cff7-4d4c-b463-8c25a7b0e48e-kube-api-access-qfhwg\") pod \"horizon-9c7f9b5b9-6878s\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.133419 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-srrcj"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.147436 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-hh6m8"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.175583 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hff8g\" (UniqueName: \"kubernetes.io/projected/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-kube-api-access-hff8g\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.175888 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-db-sync-config-data\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.176041 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.176220 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-scripts\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.176301 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.176437 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-config-data\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.176531 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-combined-ca-bundle\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.176647 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-log-httpd\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.176751 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx6tl\" (UniqueName: \"kubernetes.io/projected/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-kube-api-access-fx6tl\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.176894 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-run-httpd\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.179243 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-run-httpd\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.182327 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-log-httpd\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.182613 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.183930 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-scripts\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.184255 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-config-data\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.184921 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b54f6d777-lnnzf"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.184967 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.186186 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.205906 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.223015 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx6tl\" (UniqueName: \"kubernetes.io/projected/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-kube-api-access-fx6tl\") pod \"ceilometer-0\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.223381 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b54f6d777-lnnzf"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.239829 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-4zq28"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.241289 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.252193 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.261409 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-4zq28"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.287504 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-combined-ca-bundle\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.287575 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-config-data\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.287603 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-combined-ca-bundle\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.287646 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-scripts\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.287709 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hff8g\" (UniqueName: \"kubernetes.io/projected/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-kube-api-access-hff8g\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.287734 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-db-sync-config-data\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.287769 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2f4ea76-03f4-49e2-99d8-af57dabb6372-logs\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.287800 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw87m\" (UniqueName: \"kubernetes.io/projected/d2f4ea76-03f4-49e2-99d8-af57dabb6372-kube-api-access-bw87m\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.294231 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.306518 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-combined-ca-bundle\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.317017 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-db-sync-config-data\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.330988 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hff8g\" (UniqueName: \"kubernetes.io/projected/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-kube-api-access-hff8g\") pod \"barbican-db-sync-srrcj\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392305 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-scripts\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392580 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-logs\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392613 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392657 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrw22\" (UniqueName: \"kubernetes.io/projected/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-kube-api-access-wrw22\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392677 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7cfs\" (UniqueName: \"kubernetes.io/projected/047d386f-a54b-4cd6-a3a5-764f65f80624-kube-api-access-f7cfs\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392748 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392767 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-config\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392831 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-horizon-secret-key\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392848 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2f4ea76-03f4-49e2-99d8-af57dabb6372-logs\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.392908 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw87m\" (UniqueName: \"kubernetes.io/projected/d2f4ea76-03f4-49e2-99d8-af57dabb6372-kube-api-access-bw87m\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.393033 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-scripts\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.393103 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.393163 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-config-data\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.393187 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2f4ea76-03f4-49e2-99d8-af57dabb6372-logs\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.393216 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.393254 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-config-data\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.393512 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-combined-ca-bundle\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.397054 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-config-data\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.397281 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-scripts\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.402138 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-combined-ca-bundle\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.414922 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-srrcj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.419797 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw87m\" (UniqueName: \"kubernetes.io/projected/d2f4ea76-03f4-49e2-99d8-af57dabb6372-kube-api-access-bw87m\") pod \"placement-db-sync-5578x\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.450651 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5578x" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498464 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-config-data\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498523 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498575 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-logs\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498599 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498620 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrw22\" (UniqueName: \"kubernetes.io/projected/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-kube-api-access-wrw22\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498644 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7cfs\" (UniqueName: \"kubernetes.io/projected/047d386f-a54b-4cd6-a3a5-764f65f80624-kube-api-access-f7cfs\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498699 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498715 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-config\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498780 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-horizon-secret-key\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498856 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-scripts\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.498888 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.500492 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.502337 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-config-data\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.502439 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.502968 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-config\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.504208 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-scripts\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.504217 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.504501 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.504969 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-logs\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.510608 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-horizon-secret-key\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.534143 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrw22\" (UniqueName: \"kubernetes.io/projected/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-kube-api-access-wrw22\") pod \"horizon-5b54f6d777-lnnzf\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.540347 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7cfs\" (UniqueName: \"kubernetes.io/projected/047d386f-a54b-4cd6-a3a5-764f65f80624-kube-api-access-f7cfs\") pod \"dnsmasq-dns-76fcf4b695-4zq28\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.547482 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.715683 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4vnm9"] Oct 04 04:53:01 crc kubenswrapper[4992]: W1004 04:53:01.718255 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae4df36e_3679_40cb_9cbf_3cb8e15a5a3d.slice/crio-57324a888b711f473e85f40e3b0f483a1c514a662210f73f6ee34c6a33167552 WatchSource:0}: Error finding container 57324a888b711f473e85f40e3b0f483a1c514a662210f73f6ee34c6a33167552: Status 404 returned error can't find the container with id 57324a888b711f473e85f40e3b0f483a1c514a662210f73f6ee34c6a33167552 Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.786866 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.853992 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-hh6m8"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.888026 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-9pjhh"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.893272 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.895964 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.896195 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.898728 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-czbnj" Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.919161 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9pjhh"] Oct 04 04:53:01 crc kubenswrapper[4992]: I1004 04:53:01.992432 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-zcz9v"] Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.007573 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-combined-ca-bundle\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.007708 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-config\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.007783 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7bf7\" (UniqueName: \"kubernetes.io/projected/eb31ac75-c92d-4b64-87b0-b5dfd194617b-kube-api-access-c7bf7\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.013487 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.021926 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9c7f9b5b9-6878s"] Oct 04 04:53:02 crc kubenswrapper[4992]: W1004 04:53:02.028594 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a96c84d_f51b_48c0_9fa8_6040ed4f40f6.slice/crio-5b5568e537b1f23ebf4333c8dd7fa7944f5f2211b6e3965938bfe856d99094c0 WatchSource:0}: Error finding container 5b5568e537b1f23ebf4333c8dd7fa7944f5f2211b6e3965938bfe856d99094c0: Status 404 returned error can't find the container with id 5b5568e537b1f23ebf4333c8dd7fa7944f5f2211b6e3965938bfe856d99094c0 Oct 04 04:53:02 crc kubenswrapper[4992]: W1004 04:53:02.031096 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76056377_cff7_4d4c_b463_8c25a7b0e48e.slice/crio-0c25c6bb99f24d3df9f871a8c9dbba895d4d97c6891fd0192b39d81e0b00f512 WatchSource:0}: Error finding container 0c25c6bb99f24d3df9f871a8c9dbba895d4d97c6891fd0192b39d81e0b00f512: Status 404 returned error can't find the container with id 0c25c6bb99f24d3df9f871a8c9dbba895d4d97c6891fd0192b39d81e0b00f512 Oct 04 04:53:02 crc kubenswrapper[4992]: W1004 04:53:02.042668 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e8b3ef0_9f96_4b4b_bb9c_bcab583a8652.slice/crio-a10db9c31065f3b54aa484e7e6de2bd45aa81e67f7b56852c939936fd3336fcc WatchSource:0}: Error finding container a10db9c31065f3b54aa484e7e6de2bd45aa81e67f7b56852c939936fd3336fcc: Status 404 returned error can't find the container with id a10db9c31065f3b54aa484e7e6de2bd45aa81e67f7b56852c939936fd3336fcc Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.109764 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-config\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.109818 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7bf7\" (UniqueName: \"kubernetes.io/projected/eb31ac75-c92d-4b64-87b0-b5dfd194617b-kube-api-access-c7bf7\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.109920 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-combined-ca-bundle\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.119927 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-config\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.120338 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-combined-ca-bundle\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.159186 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7bf7\" (UniqueName: \"kubernetes.io/projected/eb31ac75-c92d-4b64-87b0-b5dfd194617b-kube-api-access-c7bf7\") pod \"neutron-db-sync-9pjhh\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.207734 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-srrcj"] Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.224670 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.246122 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5578x"] Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.266679 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-4zq28"] Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.284327 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b54f6d777-lnnzf"] Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.383529 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4vnm9" event={"ID":"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d","Type":"ContainerStarted","Data":"0eb64e942f7db33e5f76215eb6d4c90ca70cfc664afcfd27a8ac2410f58e2fed"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.383806 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4vnm9" event={"ID":"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d","Type":"ContainerStarted","Data":"57324a888b711f473e85f40e3b0f483a1c514a662210f73f6ee34c6a33167552"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.389572 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" event={"ID":"047d386f-a54b-4cd6-a3a5-764f65f80624","Type":"ContainerStarted","Data":"79797dabdd8a1b294780bcad6e4d85cb76c953da37e8828a9927e5a03b445a6a"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.406075 4992 generic.go:334] "Generic (PLEG): container finished" podID="aaa6ec0a-ce61-49e1-9e4b-78639257598f" containerID="3b4713cb9e276a59b02c8b59b3bc0e367178609eece8f104cab59b78db6625fb" exitCode=0 Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.406281 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" event={"ID":"aaa6ec0a-ce61-49e1-9e4b-78639257598f","Type":"ContainerDied","Data":"3b4713cb9e276a59b02c8b59b3bc0e367178609eece8f104cab59b78db6625fb"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.406315 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" event={"ID":"aaa6ec0a-ce61-49e1-9e4b-78639257598f","Type":"ContainerStarted","Data":"ecd3d1eadd07222c8f26c6e8d801402cb72a8f8a7f36a3d4ce341aa97be28e1d"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.413586 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b54f6d777-lnnzf" event={"ID":"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e","Type":"ContainerStarted","Data":"00559dbab726957e80d180f9b64809be91ef420d3130f3a86c873dcb3c877110"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.416121 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-srrcj" event={"ID":"15d2bb9c-fea2-48e2-aac2-3f07c2813b25","Type":"ContainerStarted","Data":"aeb93c0d887d6a84f3128995b7c0a4b5e6e555fd8e148dd83f0d0061e70b21bc"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.416247 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4vnm9" podStartSLOduration=2.416234531 podStartE2EDuration="2.416234531s" podCreationTimestamp="2025-10-04 04:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:02.404471038 +0000 UTC m=+1216.252146526" watchObservedRunningTime="2025-10-04 04:53:02.416234531 +0000 UTC m=+1216.263909999" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.417754 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zcz9v" event={"ID":"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6","Type":"ContainerStarted","Data":"5b5568e537b1f23ebf4333c8dd7fa7944f5f2211b6e3965938bfe856d99094c0"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.418865 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652","Type":"ContainerStarted","Data":"a10db9c31065f3b54aa484e7e6de2bd45aa81e67f7b56852c939936fd3336fcc"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.420275 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5578x" event={"ID":"d2f4ea76-03f4-49e2-99d8-af57dabb6372","Type":"ContainerStarted","Data":"776eef29845ab66c45342c8f52026f7c08373b9e046b88a644833112f4817939"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.424946 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" podUID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" containerName="dnsmasq-dns" containerID="cri-o://1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63" gracePeriod=10 Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.425275 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c7f9b5b9-6878s" event={"ID":"76056377-cff7-4d4c-b463-8c25a7b0e48e","Type":"ContainerStarted","Data":"0c25c6bb99f24d3df9f871a8c9dbba895d4d97c6891fd0192b39d81e0b00f512"} Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.751317 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-9pjhh"] Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.906839 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:02 crc kubenswrapper[4992]: I1004 04:53:02.939332 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.047268 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-sb\") pod \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.047343 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-svc\") pod \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.047480 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-svc\") pod \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.047552 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-config\") pod \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.047585 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-sb\") pod \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.047634 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb7kl\" (UniqueName: \"kubernetes.io/projected/aaa6ec0a-ce61-49e1-9e4b-78639257598f-kube-api-access-zb7kl\") pod \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.047660 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-swift-storage-0\") pod \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.049409 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-swift-storage-0\") pod \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.049917 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-nb\") pod \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.050278 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gqxg\" (UniqueName: \"kubernetes.io/projected/84d13164-e16f-4a6a-b8a0-5bcba00d4649-kube-api-access-4gqxg\") pod \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\" (UID: \"84d13164-e16f-4a6a-b8a0-5bcba00d4649\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.050478 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-nb\") pod \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.050619 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-config\") pod \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\" (UID: \"aaa6ec0a-ce61-49e1-9e4b-78639257598f\") " Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.080993 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d13164-e16f-4a6a-b8a0-5bcba00d4649-kube-api-access-4gqxg" (OuterVolumeSpecName: "kube-api-access-4gqxg") pod "84d13164-e16f-4a6a-b8a0-5bcba00d4649" (UID: "84d13164-e16f-4a6a-b8a0-5bcba00d4649"). InnerVolumeSpecName "kube-api-access-4gqxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.106401 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaa6ec0a-ce61-49e1-9e4b-78639257598f-kube-api-access-zb7kl" (OuterVolumeSpecName: "kube-api-access-zb7kl") pod "aaa6ec0a-ce61-49e1-9e4b-78639257598f" (UID: "aaa6ec0a-ce61-49e1-9e4b-78639257598f"). InnerVolumeSpecName "kube-api-access-zb7kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.155326 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gqxg\" (UniqueName: \"kubernetes.io/projected/84d13164-e16f-4a6a-b8a0-5bcba00d4649-kube-api-access-4gqxg\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.155353 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb7kl\" (UniqueName: \"kubernetes.io/projected/aaa6ec0a-ce61-49e1-9e4b-78639257598f-kube-api-access-zb7kl\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.179640 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9c7f9b5b9-6878s"] Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.209985 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "84d13164-e16f-4a6a-b8a0-5bcba00d4649" (UID: "84d13164-e16f-4a6a-b8a0-5bcba00d4649"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.240645 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c6dc74865-2rgnq"] Oct 04 04:53:03 crc kubenswrapper[4992]: E1004 04:53:03.244930 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" containerName="dnsmasq-dns" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.244958 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" containerName="dnsmasq-dns" Oct 04 04:53:03 crc kubenswrapper[4992]: E1004 04:53:03.244972 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" containerName="init" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.244978 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" containerName="init" Oct 04 04:53:03 crc kubenswrapper[4992]: E1004 04:53:03.245012 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaa6ec0a-ce61-49e1-9e4b-78639257598f" containerName="init" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.245019 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaa6ec0a-ce61-49e1-9e4b-78639257598f" containerName="init" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.245192 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaa6ec0a-ce61-49e1-9e4b-78639257598f" containerName="init" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.245214 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" containerName="dnsmasq-dns" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.256403 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "84d13164-e16f-4a6a-b8a0-5bcba00d4649" (UID: "84d13164-e16f-4a6a-b8a0-5bcba00d4649"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.256894 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.256922 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.260503 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aaa6ec0a-ce61-49e1-9e4b-78639257598f" (UID: "aaa6ec0a-ce61-49e1-9e4b-78639257598f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.266724 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c6dc74865-2rgnq"] Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.267274 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.277687 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aaa6ec0a-ce61-49e1-9e4b-78639257598f" (UID: "aaa6ec0a-ce61-49e1-9e4b-78639257598f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.293217 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-config" (OuterVolumeSpecName: "config") pod "84d13164-e16f-4a6a-b8a0-5bcba00d4649" (UID: "84d13164-e16f-4a6a-b8a0-5bcba00d4649"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.299480 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aaa6ec0a-ce61-49e1-9e4b-78639257598f" (UID: "aaa6ec0a-ce61-49e1-9e4b-78639257598f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.312758 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aaa6ec0a-ce61-49e1-9e4b-78639257598f" (UID: "aaa6ec0a-ce61-49e1-9e4b-78639257598f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.313235 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.313454 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-config" (OuterVolumeSpecName: "config") pod "aaa6ec0a-ce61-49e1-9e4b-78639257598f" (UID: "aaa6ec0a-ce61-49e1-9e4b-78639257598f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.350649 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "84d13164-e16f-4a6a-b8a0-5bcba00d4649" (UID: "84d13164-e16f-4a6a-b8a0-5bcba00d4649"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.358467 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c198d3-0808-4077-951a-fe14a94409ea-logs\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.358540 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6f8c\" (UniqueName: \"kubernetes.io/projected/94c198d3-0808-4077-951a-fe14a94409ea-kube-api-access-g6f8c\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.358960 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94c198d3-0808-4077-951a-fe14a94409ea-horizon-secret-key\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359034 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-scripts\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359069 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-config-data\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359149 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359166 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359178 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359191 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359203 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359214 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.359227 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aaa6ec0a-ce61-49e1-9e4b-78639257598f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.390693 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "84d13164-e16f-4a6a-b8a0-5bcba00d4649" (UID: "84d13164-e16f-4a6a-b8a0-5bcba00d4649"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.451684 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.451736 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-hh6m8" event={"ID":"aaa6ec0a-ce61-49e1-9e4b-78639257598f","Type":"ContainerDied","Data":"ecd3d1eadd07222c8f26c6e8d801402cb72a8f8a7f36a3d4ce341aa97be28e1d"} Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.451805 4992 scope.go:117] "RemoveContainer" containerID="3b4713cb9e276a59b02c8b59b3bc0e367178609eece8f104cab59b78db6625fb" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.460626 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94c198d3-0808-4077-951a-fe14a94409ea-horizon-secret-key\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.460687 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-scripts\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.460723 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-config-data\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.460823 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c198d3-0808-4077-951a-fe14a94409ea-logs\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.460865 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6f8c\" (UniqueName: \"kubernetes.io/projected/94c198d3-0808-4077-951a-fe14a94409ea-kube-api-access-g6f8c\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.460924 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/84d13164-e16f-4a6a-b8a0-5bcba00d4649-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.462033 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-scripts\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.462916 4992 generic.go:334] "Generic (PLEG): container finished" podID="2ea89995-b28e-471b-8d9e-6f97d93fe073" containerID="33de08d820fa120fa3e4c06b5ec746bda80c31ac74728f3c9c351b9bf084588c" exitCode=0 Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.463049 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-chkks" event={"ID":"2ea89995-b28e-471b-8d9e-6f97d93fe073","Type":"ContainerDied","Data":"33de08d820fa120fa3e4c06b5ec746bda80c31ac74728f3c9c351b9bf084588c"} Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.463619 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-config-data\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.463948 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c198d3-0808-4077-951a-fe14a94409ea-logs\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.475174 4992 generic.go:334] "Generic (PLEG): container finished" podID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" containerID="1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63" exitCode=0 Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.475316 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" event={"ID":"84d13164-e16f-4a6a-b8a0-5bcba00d4649","Type":"ContainerDied","Data":"1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63"} Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.475402 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" event={"ID":"84d13164-e16f-4a6a-b8a0-5bcba00d4649","Type":"ContainerDied","Data":"05b08aa266af718f2121b56d26f31c2304129d43b710ce460d40ae1b8c84768a"} Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.475413 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94c198d3-0808-4077-951a-fe14a94409ea-horizon-secret-key\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.475503 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-x4zbx" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.484905 4992 generic.go:334] "Generic (PLEG): container finished" podID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerID="ef59134edd92d7f19b87db1c6e9a2d33fe39b9e0d71a951f4b058b108ed71b13" exitCode=0 Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.484986 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" event={"ID":"047d386f-a54b-4cd6-a3a5-764f65f80624","Type":"ContainerDied","Data":"ef59134edd92d7f19b87db1c6e9a2d33fe39b9e0d71a951f4b058b108ed71b13"} Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.494121 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9pjhh" event={"ID":"eb31ac75-c92d-4b64-87b0-b5dfd194617b","Type":"ContainerStarted","Data":"d463686a8b7b1c4500d12116d647ebe0456ddb474581d901a3f5e2ea0ea04ccd"} Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.494751 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9pjhh" event={"ID":"eb31ac75-c92d-4b64-87b0-b5dfd194617b","Type":"ContainerStarted","Data":"dc8af903ef9e979b6f97cbce8cbe03feb7430493900c97a0703b8a3c5a2bb08f"} Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.495231 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6f8c\" (UniqueName: \"kubernetes.io/projected/94c198d3-0808-4077-951a-fe14a94409ea-kube-api-access-g6f8c\") pod \"horizon-c6dc74865-2rgnq\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.542871 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-9pjhh" podStartSLOduration=2.542798086 podStartE2EDuration="2.542798086s" podCreationTimestamp="2025-10-04 04:53:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:03.525925036 +0000 UTC m=+1217.373600504" watchObservedRunningTime="2025-10-04 04:53:03.542798086 +0000 UTC m=+1217.390473554" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.587881 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x4zbx"] Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.591629 4992 scope.go:117] "RemoveContainer" containerID="1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.599734 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-x4zbx"] Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.621832 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-hh6m8"] Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.628668 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-hh6m8"] Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.681539 4992 scope.go:117] "RemoveContainer" containerID="00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.732580 4992 scope.go:117] "RemoveContainer" containerID="1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63" Oct 04 04:53:03 crc kubenswrapper[4992]: E1004 04:53:03.733927 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63\": container with ID starting with 1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63 not found: ID does not exist" containerID="1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.733976 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63"} err="failed to get container status \"1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63\": rpc error: code = NotFound desc = could not find container \"1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63\": container with ID starting with 1f8ee7fcb4f22ee3afb174d0735769b80bf575404d34323ebcb8fc845b94bb63 not found: ID does not exist" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.734002 4992 scope.go:117] "RemoveContainer" containerID="00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602" Oct 04 04:53:03 crc kubenswrapper[4992]: E1004 04:53:03.736792 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602\": container with ID starting with 00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602 not found: ID does not exist" containerID="00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.736846 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602"} err="failed to get container status \"00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602\": rpc error: code = NotFound desc = could not find container \"00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602\": container with ID starting with 00d5a54dcf802273a82889f2f40790e870078317f0eabb04c5aa1ee9beeba602 not found: ID does not exist" Oct 04 04:53:03 crc kubenswrapper[4992]: I1004 04:53:03.750915 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:04 crc kubenswrapper[4992]: I1004 04:53:04.306913 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c6dc74865-2rgnq"] Oct 04 04:53:04 crc kubenswrapper[4992]: I1004 04:53:04.367874 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d13164-e16f-4a6a-b8a0-5bcba00d4649" path="/var/lib/kubelet/pods/84d13164-e16f-4a6a-b8a0-5bcba00d4649/volumes" Oct 04 04:53:04 crc kubenswrapper[4992]: I1004 04:53:04.371172 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaa6ec0a-ce61-49e1-9e4b-78639257598f" path="/var/lib/kubelet/pods/aaa6ec0a-ce61-49e1-9e4b-78639257598f/volumes" Oct 04 04:53:04 crc kubenswrapper[4992]: I1004 04:53:04.515622 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6dc74865-2rgnq" event={"ID":"94c198d3-0808-4077-951a-fe14a94409ea","Type":"ContainerStarted","Data":"2523dae3be6d8dd6670c2ec26687a5726194a8d297a671c1803af2a5938e33f6"} Oct 04 04:53:04 crc kubenswrapper[4992]: I1004 04:53:04.527384 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" event={"ID":"047d386f-a54b-4cd6-a3a5-764f65f80624","Type":"ContainerStarted","Data":"cc8a2ece388265fee5ddbca6f1dae5aac230814a43fb5574481ad6829fde08e0"} Oct 04 04:53:04 crc kubenswrapper[4992]: I1004 04:53:04.528281 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:04 crc kubenswrapper[4992]: I1004 04:53:04.558201 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" podStartSLOduration=3.5581806780000003 podStartE2EDuration="3.558180678s" podCreationTimestamp="2025-10-04 04:53:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:04.546776494 +0000 UTC m=+1218.394451962" watchObservedRunningTime="2025-10-04 04:53:04.558180678 +0000 UTC m=+1218.405856146" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.308136 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-chkks" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.399415 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-db-sync-config-data\") pod \"2ea89995-b28e-471b-8d9e-6f97d93fe073\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.399532 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-config-data\") pod \"2ea89995-b28e-471b-8d9e-6f97d93fe073\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.399554 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dzmk\" (UniqueName: \"kubernetes.io/projected/2ea89995-b28e-471b-8d9e-6f97d93fe073-kube-api-access-9dzmk\") pod \"2ea89995-b28e-471b-8d9e-6f97d93fe073\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.399579 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-combined-ca-bundle\") pod \"2ea89995-b28e-471b-8d9e-6f97d93fe073\" (UID: \"2ea89995-b28e-471b-8d9e-6f97d93fe073\") " Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.406641 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea89995-b28e-471b-8d9e-6f97d93fe073-kube-api-access-9dzmk" (OuterVolumeSpecName: "kube-api-access-9dzmk") pod "2ea89995-b28e-471b-8d9e-6f97d93fe073" (UID: "2ea89995-b28e-471b-8d9e-6f97d93fe073"). InnerVolumeSpecName "kube-api-access-9dzmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.413599 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2ea89995-b28e-471b-8d9e-6f97d93fe073" (UID: "2ea89995-b28e-471b-8d9e-6f97d93fe073"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.427267 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ea89995-b28e-471b-8d9e-6f97d93fe073" (UID: "2ea89995-b28e-471b-8d9e-6f97d93fe073"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.501673 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dzmk\" (UniqueName: \"kubernetes.io/projected/2ea89995-b28e-471b-8d9e-6f97d93fe073-kube-api-access-9dzmk\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.501707 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.501717 4992 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.505334 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-config-data" (OuterVolumeSpecName: "config-data") pod "2ea89995-b28e-471b-8d9e-6f97d93fe073" (UID: "2ea89995-b28e-471b-8d9e-6f97d93fe073"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.544443 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-chkks" event={"ID":"2ea89995-b28e-471b-8d9e-6f97d93fe073","Type":"ContainerDied","Data":"bf7d6e2d39296c47478d6f1d5c2aacfa57ee7f732643cbe38fac97e23a66c053"} Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.544486 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf7d6e2d39296c47478d6f1d5c2aacfa57ee7f732643cbe38fac97e23a66c053" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.544538 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-chkks" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.603058 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ea89995-b28e-471b-8d9e-6f97d93fe073-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.931777 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-4zq28"] Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.972665 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-qs2sd"] Oct 04 04:53:05 crc kubenswrapper[4992]: E1004 04:53:05.973111 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea89995-b28e-471b-8d9e-6f97d93fe073" containerName="glance-db-sync" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.973127 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea89995-b28e-471b-8d9e-6f97d93fe073" containerName="glance-db-sync" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.973383 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea89995-b28e-471b-8d9e-6f97d93fe073" containerName="glance-db-sync" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.974484 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:05 crc kubenswrapper[4992]: I1004 04:53:05.995221 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-qs2sd"] Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.128437 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.128535 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-config\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.128590 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.128631 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.128662 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76x4g\" (UniqueName: \"kubernetes.io/projected/8e8eb649-79cf-41a7-a2be-a0d713c6d741-kube-api-access-76x4g\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.128684 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.232030 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.232342 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.232395 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76x4g\" (UniqueName: \"kubernetes.io/projected/8e8eb649-79cf-41a7-a2be-a0d713c6d741-kube-api-access-76x4g\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.232415 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.232460 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.232528 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-config\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.233165 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.233283 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-config\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.233902 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.234177 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.234547 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.256935 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76x4g\" (UniqueName: \"kubernetes.io/projected/8e8eb649-79cf-41a7-a2be-a0d713c6d741-kube-api-access-76x4g\") pod \"dnsmasq-dns-8b5c85b87-qs2sd\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.305907 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.554953 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" containerID="cri-o://cc8a2ece388265fee5ddbca6f1dae5aac230814a43fb5574481ad6829fde08e0" gracePeriod=10 Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.810105 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.812772 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.820203 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.820561 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zjhxk" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.820847 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.825604 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-qs2sd"] Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.835398 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.953042 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-logs\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.953120 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-scripts\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.953145 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-config-data\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.953197 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.953225 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cdrg\" (UniqueName: \"kubernetes.io/projected/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-kube-api-access-8cdrg\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.953256 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:06 crc kubenswrapper[4992]: I1004 04:53:06.953297 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.054716 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cdrg\" (UniqueName: \"kubernetes.io/projected/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-kube-api-access-8cdrg\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.054776 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.054796 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.054894 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-logs\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.054941 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-scripts\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.054964 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-config-data\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.055008 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.055428 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.056029 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.056186 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-logs\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.060827 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-scripts\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.061397 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-config-data\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.065761 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.073960 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cdrg\" (UniqueName: \"kubernetes.io/projected/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-kube-api-access-8cdrg\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.102187 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.102709 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.109153 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.111322 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.113729 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.145022 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.258193 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.258303 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.258352 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.258466 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9f88\" (UniqueName: \"kubernetes.io/projected/01f98839-df68-4910-9a54-2236f5f1e6e6-kube-api-access-p9f88\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.258493 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.258530 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.258598 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.360807 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.360873 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.360908 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.360977 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9f88\" (UniqueName: \"kubernetes.io/projected/01f98839-df68-4910-9a54-2236f5f1e6e6-kube-api-access-p9f88\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.361001 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.361030 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.361085 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.361205 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.361631 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.361700 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.369051 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.369430 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.382129 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9f88\" (UniqueName: \"kubernetes.io/projected/01f98839-df68-4910-9a54-2236f5f1e6e6-kube-api-access-p9f88\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.386523 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.401973 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.485434 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.568836 4992 generic.go:334] "Generic (PLEG): container finished" podID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerID="cc8a2ece388265fee5ddbca6f1dae5aac230814a43fb5574481ad6829fde08e0" exitCode=0 Oct 04 04:53:07 crc kubenswrapper[4992]: I1004 04:53:07.568886 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" event={"ID":"047d386f-a54b-4cd6-a3a5-764f65f80624","Type":"ContainerDied","Data":"cc8a2ece388265fee5ddbca6f1dae5aac230814a43fb5574481ad6829fde08e0"} Oct 04 04:53:08 crc kubenswrapper[4992]: I1004 04:53:08.592254 4992 generic.go:334] "Generic (PLEG): container finished" podID="ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" containerID="0eb64e942f7db33e5f76215eb6d4c90ca70cfc664afcfd27a8ac2410f58e2fed" exitCode=0 Oct 04 04:53:08 crc kubenswrapper[4992]: I1004 04:53:08.592338 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4vnm9" event={"ID":"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d","Type":"ContainerDied","Data":"0eb64e942f7db33e5f76215eb6d4c90ca70cfc664afcfd27a8ac2410f58e2fed"} Oct 04 04:53:11 crc kubenswrapper[4992]: I1004 04:53:11.790509 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Oct 04 04:53:12 crc kubenswrapper[4992]: I1004 04:53:12.674841 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:12 crc kubenswrapper[4992]: I1004 04:53:12.879825 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.503297 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b54f6d777-lnnzf"] Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.526668 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.573816 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-combined-ca-bundle\") pod \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.573901 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-fernet-keys\") pod \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.573964 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddmjt\" (UniqueName: \"kubernetes.io/projected/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-kube-api-access-ddmjt\") pod \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.574009 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-scripts\") pod \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.574076 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-config-data\") pod \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.574208 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-credential-keys\") pod \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\" (UID: \"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d\") " Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.595612 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-55b5954cd6-68rnc"] Oct 04 04:53:13 crc kubenswrapper[4992]: E1004 04:53:13.596083 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" containerName="keystone-bootstrap" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.596108 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" containerName="keystone-bootstrap" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.596353 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" containerName="keystone-bootstrap" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.597587 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.604465 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-scripts" (OuterVolumeSpecName: "scripts") pod "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" (UID: "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.604903 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" (UID: "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.605005 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.605573 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-kube-api-access-ddmjt" (OuterVolumeSpecName: "kube-api-access-ddmjt") pod "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" (UID: "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d"). InnerVolumeSpecName "kube-api-access-ddmjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.609686 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" (UID: "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.611980 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55b5954cd6-68rnc"] Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.623597 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" (UID: "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.658955 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" event={"ID":"8e8eb649-79cf-41a7-a2be-a0d713c6d741","Type":"ContainerStarted","Data":"ff4ae11e26c176575c66a03fbaeccd29d82a31185804d7a38f920b41b643aaa2"} Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.663766 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4vnm9" event={"ID":"ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d","Type":"ContainerDied","Data":"57324a888b711f473e85f40e3b0f483a1c514a662210f73f6ee34c6a33167552"} Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.663818 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57324a888b711f473e85f40e3b0f483a1c514a662210f73f6ee34c6a33167552" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.663906 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4vnm9" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.676174 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwwvq\" (UniqueName: \"kubernetes.io/projected/d75279dd-898a-4ffe-85e6-804f066de711-kube-api-access-fwwvq\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.676279 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-horizon-secret-key\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.676323 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-horizon-tls-certs\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.676423 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d75279dd-898a-4ffe-85e6-804f066de711-scripts\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.676448 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d75279dd-898a-4ffe-85e6-804f066de711-config-data\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.676810 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-combined-ca-bundle\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.676883 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d75279dd-898a-4ffe-85e6-804f066de711-logs\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.677144 4992 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.677158 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.677172 4992 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.677184 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddmjt\" (UniqueName: \"kubernetes.io/projected/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-kube-api-access-ddmjt\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.677196 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.703955 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-config-data" (OuterVolumeSpecName: "config-data") pod "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" (UID: "ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.778671 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-horizon-tls-certs\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.778721 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d75279dd-898a-4ffe-85e6-804f066de711-scripts\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.778748 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d75279dd-898a-4ffe-85e6-804f066de711-config-data\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.778795 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-combined-ca-bundle\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.778817 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d75279dd-898a-4ffe-85e6-804f066de711-logs\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.778886 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwwvq\" (UniqueName: \"kubernetes.io/projected/d75279dd-898a-4ffe-85e6-804f066de711-kube-api-access-fwwvq\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.778918 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-horizon-secret-key\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.778979 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.779638 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d75279dd-898a-4ffe-85e6-804f066de711-scripts\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.779959 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d75279dd-898a-4ffe-85e6-804f066de711-logs\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.781261 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d75279dd-898a-4ffe-85e6-804f066de711-config-data\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.782613 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-horizon-secret-key\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.783842 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-horizon-tls-certs\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.796675 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d75279dd-898a-4ffe-85e6-804f066de711-combined-ca-bundle\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:13 crc kubenswrapper[4992]: I1004 04:53:13.799931 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwwvq\" (UniqueName: \"kubernetes.io/projected/d75279dd-898a-4ffe-85e6-804f066de711-kube-api-access-fwwvq\") pod \"horizon-55b5954cd6-68rnc\" (UID: \"d75279dd-898a-4ffe-85e6-804f066de711\") " pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.055297 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.506951 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4vnm9"] Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.515392 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4vnm9"] Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.609029 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ln7d7"] Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.610489 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.613047 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5nrbs" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.613252 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.613413 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.613559 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.622514 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ln7d7"] Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.694741 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-scripts\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.695022 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-combined-ca-bundle\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.695048 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-config-data\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.695094 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-credential-keys\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.695165 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-fernet-keys\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.695190 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7f9j\" (UniqueName: \"kubernetes.io/projected/dd8feb2f-147b-43d8-ab87-58703170e959-kube-api-access-t7f9j\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.796652 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-fernet-keys\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.796689 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7f9j\" (UniqueName: \"kubernetes.io/projected/dd8feb2f-147b-43d8-ab87-58703170e959-kube-api-access-t7f9j\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.796784 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-scripts\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.796814 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-combined-ca-bundle\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.796835 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-config-data\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.796879 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-credential-keys\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.800027 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-scripts\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.800212 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-combined-ca-bundle\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.801031 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-config-data\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.801416 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-credential-keys\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.803378 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-fernet-keys\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.814165 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7f9j\" (UniqueName: \"kubernetes.io/projected/dd8feb2f-147b-43d8-ab87-58703170e959-kube-api-access-t7f9j\") pod \"keystone-bootstrap-ln7d7\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:14 crc kubenswrapper[4992]: I1004 04:53:14.941834 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:16 crc kubenswrapper[4992]: I1004 04:53:16.333924 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d" path="/var/lib/kubelet/pods/ae4df36e-3679-40cb-9cbf-3cb8e15a5a3d/volumes" Oct 04 04:53:17 crc kubenswrapper[4992]: E1004 04:53:17.829983 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 04 04:53:17 crc kubenswrapper[4992]: E1004 04:53:17.830180 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bw87m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-5578x_openstack(d2f4ea76-03f4-49e2-99d8-af57dabb6372): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:53:17 crc kubenswrapper[4992]: E1004 04:53:17.831390 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-5578x" podUID="d2f4ea76-03f4-49e2-99d8-af57dabb6372" Oct 04 04:53:18 crc kubenswrapper[4992]: E1004 04:53:18.707130 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-5578x" podUID="d2f4ea76-03f4-49e2-99d8-af57dabb6372" Oct 04 04:53:20 crc kubenswrapper[4992]: I1004 04:53:20.894565 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:53:20 crc kubenswrapper[4992]: I1004 04:53:20.896061 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:53:21 crc kubenswrapper[4992]: I1004 04:53:21.787613 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Oct 04 04:53:26 crc kubenswrapper[4992]: I1004 04:53:26.788732 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Oct 04 04:53:31 crc kubenswrapper[4992]: I1004 04:53:31.789418 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.046688 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.047100 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n558hd4h574h8bh65h7bh566h5bchc8h56fh695h5c4h5bchd9h95h5dfh5b5h64bh5d4hb8hf6h678h5b7hc6h78h68dh655hdhbbhfh5dbh5d6q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qfhwg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-9c7f9b5b9-6878s_openstack(76056377-cff7-4d4c-b463-8c25a7b0e48e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.056403 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-9c7f9b5b9-6878s" podUID="76056377-cff7-4d4c-b463-8c25a7b0e48e" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.065305 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.065454 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd6h79h66fh666h598hddh567h695h95h54h5bdh594h55bh56bh4hd4h54dh54h598h699hcdh8ch59ch5cdh9bhddh5bfh85h568hb7h85h67bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wrw22,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5b54f6d777-lnnzf_openstack(4d5c9d6f-8a5c-470e-b90c-7716dc0f635e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.080342 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5b54f6d777-lnnzf" podUID="4d5c9d6f-8a5c-470e-b90c-7716dc0f635e" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.096876 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.097138 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf8h84h96h6fh9h586h56chb6hd8hc7h5f7hdfh54ch88h5bdhf4h5f7h696h64h578h599h5b7h5d6h5bfhfch5dfh658h659h55ch55fh64ch66bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g6f8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-c6dc74865-2rgnq_openstack(94c198d3-0808-4077-951a-fe14a94409ea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.099298 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-c6dc74865-2rgnq" podUID="94c198d3-0808-4077-951a-fe14a94409ea" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.143299 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.164159 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7cfs\" (UniqueName: \"kubernetes.io/projected/047d386f-a54b-4cd6-a3a5-764f65f80624-kube-api-access-f7cfs\") pod \"047d386f-a54b-4cd6-a3a5-764f65f80624\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.164308 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-nb\") pod \"047d386f-a54b-4cd6-a3a5-764f65f80624\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.164419 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-svc\") pod \"047d386f-a54b-4cd6-a3a5-764f65f80624\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.164475 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-swift-storage-0\") pod \"047d386f-a54b-4cd6-a3a5-764f65f80624\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.164500 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-config\") pod \"047d386f-a54b-4cd6-a3a5-764f65f80624\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.164516 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-sb\") pod \"047d386f-a54b-4cd6-a3a5-764f65f80624\" (UID: \"047d386f-a54b-4cd6-a3a5-764f65f80624\") " Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.183648 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/047d386f-a54b-4cd6-a3a5-764f65f80624-kube-api-access-f7cfs" (OuterVolumeSpecName: "kube-api-access-f7cfs") pod "047d386f-a54b-4cd6-a3a5-764f65f80624" (UID: "047d386f-a54b-4cd6-a3a5-764f65f80624"). InnerVolumeSpecName "kube-api-access-f7cfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.221962 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "047d386f-a54b-4cd6-a3a5-764f65f80624" (UID: "047d386f-a54b-4cd6-a3a5-764f65f80624"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.225303 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "047d386f-a54b-4cd6-a3a5-764f65f80624" (UID: "047d386f-a54b-4cd6-a3a5-764f65f80624"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.238385 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "047d386f-a54b-4cd6-a3a5-764f65f80624" (UID: "047d386f-a54b-4cd6-a3a5-764f65f80624"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.248224 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-config" (OuterVolumeSpecName: "config") pod "047d386f-a54b-4cd6-a3a5-764f65f80624" (UID: "047d386f-a54b-4cd6-a3a5-764f65f80624"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.258120 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "047d386f-a54b-4cd6-a3a5-764f65f80624" (UID: "047d386f-a54b-4cd6-a3a5-764f65f80624"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.266664 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.266707 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.266718 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.266729 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.266738 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/047d386f-a54b-4cd6-a3a5-764f65f80624-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.266746 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7cfs\" (UniqueName: \"kubernetes.io/projected/047d386f-a54b-4cd6-a3a5-764f65f80624-kube-api-access-f7cfs\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.628431 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.628642 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n54h7h56dh67dh689hc5h5fh579h88h56ch7fh658h5dfh5fh5fdh64ch55bh5f9h669h96h5f4h684h679hd6h57bh5b8hd4hf8hc4h5d5h8bh86q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fx6tl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.851048 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" event={"ID":"047d386f-a54b-4cd6-a3a5-764f65f80624","Type":"ContainerDied","Data":"79797dabdd8a1b294780bcad6e4d85cb76c953da37e8828a9927e5a03b445a6a"} Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.851116 4992 scope.go:117] "RemoveContainer" containerID="cc8a2ece388265fee5ddbca6f1dae5aac230814a43fb5574481ad6829fde08e0" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.851179 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" Oct 04 04:53:35 crc kubenswrapper[4992]: E1004 04:53:35.854091 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-c6dc74865-2rgnq" podUID="94c198d3-0808-4077-951a-fe14a94409ea" Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.944101 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-4zq28"] Oct 04 04:53:35 crc kubenswrapper[4992]: I1004 04:53:35.951070 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-4zq28"] Oct 04 04:53:36 crc kubenswrapper[4992]: I1004 04:53:36.337192 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" path="/var/lib/kubelet/pods/047d386f-a54b-4cd6-a3a5-764f65f80624/volumes" Oct 04 04:53:36 crc kubenswrapper[4992]: I1004 04:53:36.791130 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-4zq28" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.240650 4992 scope.go:117] "RemoveContainer" containerID="ef59134edd92d7f19b87db1c6e9a2d33fe39b9e0d71a951f4b058b108ed71b13" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.560039 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.587613 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610206 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-config-data\") pod \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610338 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfhwg\" (UniqueName: \"kubernetes.io/projected/76056377-cff7-4d4c-b463-8c25a7b0e48e-kube-api-access-qfhwg\") pod \"76056377-cff7-4d4c-b463-8c25a7b0e48e\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610443 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76056377-cff7-4d4c-b463-8c25a7b0e48e-horizon-secret-key\") pod \"76056377-cff7-4d4c-b463-8c25a7b0e48e\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610513 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-logs\") pod \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610549 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrw22\" (UniqueName: \"kubernetes.io/projected/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-kube-api-access-wrw22\") pod \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610580 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-config-data\") pod \"76056377-cff7-4d4c-b463-8c25a7b0e48e\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610606 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-scripts\") pod \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610628 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-scripts\") pod \"76056377-cff7-4d4c-b463-8c25a7b0e48e\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610694 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76056377-cff7-4d4c-b463-8c25a7b0e48e-logs\") pod \"76056377-cff7-4d4c-b463-8c25a7b0e48e\" (UID: \"76056377-cff7-4d4c-b463-8c25a7b0e48e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.610724 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-horizon-secret-key\") pod \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\" (UID: \"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e\") " Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.612242 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-config-data" (OuterVolumeSpecName: "config-data") pod "76056377-cff7-4d4c-b463-8c25a7b0e48e" (UID: "76056377-cff7-4d4c-b463-8c25a7b0e48e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.613194 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-scripts" (OuterVolumeSpecName: "scripts") pod "76056377-cff7-4d4c-b463-8c25a7b0e48e" (UID: "76056377-cff7-4d4c-b463-8c25a7b0e48e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.613272 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-scripts" (OuterVolumeSpecName: "scripts") pod "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e" (UID: "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.613396 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-logs" (OuterVolumeSpecName: "logs") pod "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e" (UID: "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.613495 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76056377-cff7-4d4c-b463-8c25a7b0e48e-logs" (OuterVolumeSpecName: "logs") pod "76056377-cff7-4d4c-b463-8c25a7b0e48e" (UID: "76056377-cff7-4d4c-b463-8c25a7b0e48e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.614218 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-config-data" (OuterVolumeSpecName: "config-data") pod "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e" (UID: "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.621962 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76056377-cff7-4d4c-b463-8c25a7b0e48e-kube-api-access-qfhwg" (OuterVolumeSpecName: "kube-api-access-qfhwg") pod "76056377-cff7-4d4c-b463-8c25a7b0e48e" (UID: "76056377-cff7-4d4c-b463-8c25a7b0e48e"). InnerVolumeSpecName "kube-api-access-qfhwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.623382 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76056377-cff7-4d4c-b463-8c25a7b0e48e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "76056377-cff7-4d4c-b463-8c25a7b0e48e" (UID: "76056377-cff7-4d4c-b463-8c25a7b0e48e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.625766 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e" (UID: "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.632518 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-kube-api-access-wrw22" (OuterVolumeSpecName: "kube-api-access-wrw22") pod "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e" (UID: "4d5c9d6f-8a5c-470e-b90c-7716dc0f635e"). InnerVolumeSpecName "kube-api-access-wrw22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.712930 4992 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76056377-cff7-4d4c-b463-8c25a7b0e48e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.712991 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.713010 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrw22\" (UniqueName: \"kubernetes.io/projected/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-kube-api-access-wrw22\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.713083 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.713099 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76056377-cff7-4d4c-b463-8c25a7b0e48e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.713115 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.713170 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76056377-cff7-4d4c-b463-8c25a7b0e48e-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.713186 4992 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.713201 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.713251 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfhwg\" (UniqueName: \"kubernetes.io/projected/76056377-cff7-4d4c-b463-8c25a7b0e48e-kube-api-access-qfhwg\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:37 crc kubenswrapper[4992]: E1004 04:53:37.784096 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 04 04:53:37 crc kubenswrapper[4992]: E1004 04:53:37.784596 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wpwvw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-zcz9v_openstack(0a96c84d-f51b-48c0-9fa8-6040ed4f40f6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 04:53:37 crc kubenswrapper[4992]: E1004 04:53:37.785797 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-zcz9v" podUID="0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" Oct 04 04:53:37 crc kubenswrapper[4992]: W1004 04:53:37.828458 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd75279dd_898a_4ffe_85e6_804f066de711.slice/crio-1617be38c6a014a7f6ff16ef82130a6c0bba2ba03e86e9494e5064c9384295c6 WatchSource:0}: Error finding container 1617be38c6a014a7f6ff16ef82130a6c0bba2ba03e86e9494e5064c9384295c6: Status 404 returned error can't find the container with id 1617be38c6a014a7f6ff16ef82130a6c0bba2ba03e86e9494e5064c9384295c6 Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.829989 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55b5954cd6-68rnc"] Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.865933 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c7f9b5b9-6878s" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.865958 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c7f9b5b9-6878s" event={"ID":"76056377-cff7-4d4c-b463-8c25a7b0e48e","Type":"ContainerDied","Data":"0c25c6bb99f24d3df9f871a8c9dbba895d4d97c6891fd0192b39d81e0b00f512"} Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.867532 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b54f6d777-lnnzf" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.867524 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b54f6d777-lnnzf" event={"ID":"4d5c9d6f-8a5c-470e-b90c-7716dc0f635e","Type":"ContainerDied","Data":"00559dbab726957e80d180f9b64809be91ef420d3130f3a86c873dcb3c877110"} Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.880648 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55b5954cd6-68rnc" event={"ID":"d75279dd-898a-4ffe-85e6-804f066de711","Type":"ContainerStarted","Data":"1617be38c6a014a7f6ff16ef82130a6c0bba2ba03e86e9494e5064c9384295c6"} Oct 04 04:53:37 crc kubenswrapper[4992]: E1004 04:53:37.881291 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-zcz9v" podUID="0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.938445 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ln7d7"] Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.964632 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b54f6d777-lnnzf"] Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.971524 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5b54f6d777-lnnzf"] Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.983871 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9c7f9b5b9-6878s"] Oct 04 04:53:37 crc kubenswrapper[4992]: I1004 04:53:37.990042 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-9c7f9b5b9-6878s"] Oct 04 04:53:38 crc kubenswrapper[4992]: I1004 04:53:38.328259 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d5c9d6f-8a5c-470e-b90c-7716dc0f635e" path="/var/lib/kubelet/pods/4d5c9d6f-8a5c-470e-b90c-7716dc0f635e/volumes" Oct 04 04:53:38 crc kubenswrapper[4992]: I1004 04:53:38.328712 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76056377-cff7-4d4c-b463-8c25a7b0e48e" path="/var/lib/kubelet/pods/76056377-cff7-4d4c-b463-8c25a7b0e48e/volumes" Oct 04 04:53:38 crc kubenswrapper[4992]: I1004 04:53:38.891247 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-srrcj" event={"ID":"15d2bb9c-fea2-48e2-aac2-3f07c2813b25","Type":"ContainerStarted","Data":"a708ae19fdcb6154dfbe928d522052f3b71af3f67a4382c3650b02f3c29e2a7c"} Oct 04 04:53:38 crc kubenswrapper[4992]: I1004 04:53:38.893665 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ln7d7" event={"ID":"dd8feb2f-147b-43d8-ab87-58703170e959","Type":"ContainerStarted","Data":"debf3cba486bfdfb01da926716ae7e614f2f71ca37382b9cabbcb1fb44e12bb8"} Oct 04 04:53:38 crc kubenswrapper[4992]: I1004 04:53:38.893714 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ln7d7" event={"ID":"dd8feb2f-147b-43d8-ab87-58703170e959","Type":"ContainerStarted","Data":"f01ebf0845ae4d1e15838fe6610586e68ef7291102974cae4ede345e451df768"} Oct 04 04:53:38 crc kubenswrapper[4992]: I1004 04:53:38.895294 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" event={"ID":"8e8eb649-79cf-41a7-a2be-a0d713c6d741","Type":"ContainerStarted","Data":"b51c0840455bb7516008bf91067006436cbf57cc556eeedf245977b73854a864"} Oct 04 04:53:39 crc kubenswrapper[4992]: I1004 04:53:39.911285 4992 generic.go:334] "Generic (PLEG): container finished" podID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerID="b51c0840455bb7516008bf91067006436cbf57cc556eeedf245977b73854a864" exitCode=0 Oct 04 04:53:39 crc kubenswrapper[4992]: I1004 04:53:39.911349 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" event={"ID":"8e8eb649-79cf-41a7-a2be-a0d713c6d741","Type":"ContainerDied","Data":"b51c0840455bb7516008bf91067006436cbf57cc556eeedf245977b73854a864"} Oct 04 04:53:39 crc kubenswrapper[4992]: I1004 04:53:39.925227 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:39 crc kubenswrapper[4992]: I1004 04:53:39.987348 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ln7d7" podStartSLOduration=25.987332302 podStartE2EDuration="25.987332302s" podCreationTimestamp="2025-10-04 04:53:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:39.965943243 +0000 UTC m=+1253.813618721" watchObservedRunningTime="2025-10-04 04:53:39.987332302 +0000 UTC m=+1253.835007760" Oct 04 04:53:39 crc kubenswrapper[4992]: I1004 04:53:39.993489 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-srrcj" podStartSLOduration=5.008483626 podStartE2EDuration="39.993474455s" podCreationTimestamp="2025-10-04 04:53:00 +0000 UTC" firstStartedPulling="2025-10-04 04:53:02.21810125 +0000 UTC m=+1216.065776718" lastFinishedPulling="2025-10-04 04:53:37.203092079 +0000 UTC m=+1251.050767547" observedRunningTime="2025-10-04 04:53:39.984722672 +0000 UTC m=+1253.832398140" watchObservedRunningTime="2025-10-04 04:53:39.993474455 +0000 UTC m=+1253.841149923" Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.454100 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.924887 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5578x" event={"ID":"d2f4ea76-03f4-49e2-99d8-af57dabb6372","Type":"ContainerStarted","Data":"37d9da7f7b13e8017ebf021988b1d3a87fffa72d2f31f182996f8348347bd403"} Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.927547 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a","Type":"ContainerStarted","Data":"9cabb92cf0277db5caf1115033e45f70c82d6c337d20a07f104e86d0991406ca"} Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.932374 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01f98839-df68-4910-9a54-2236f5f1e6e6","Type":"ContainerStarted","Data":"82da68eb686e1cb14a3ace3813583a92e7028695d961ea913f66b49352c4f75e"} Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.932443 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01f98839-df68-4910-9a54-2236f5f1e6e6","Type":"ContainerStarted","Data":"2e8d9e183b446633cf3082f00cc8cd70a0fadaf1f3834b6f771b8241dfad44be"} Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.934657 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55b5954cd6-68rnc" event={"ID":"d75279dd-898a-4ffe-85e6-804f066de711","Type":"ContainerStarted","Data":"7c195c82608d9590e772b8a14a93c352148e89f9c724fd244dc0ac947280bb46"} Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.934686 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55b5954cd6-68rnc" event={"ID":"d75279dd-898a-4ffe-85e6-804f066de711","Type":"ContainerStarted","Data":"63179b82e237ee4b33ae5dda11c5337436467ef995a3f3f364455c0dd5a6252e"} Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.937202 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" event={"ID":"8e8eb649-79cf-41a7-a2be-a0d713c6d741","Type":"ContainerStarted","Data":"48780c5b8756483e0bceadf4f9d0c6c2d167eade4004c20641ca7c79af35cbb5"} Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.937603 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.942129 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652","Type":"ContainerStarted","Data":"301352d0463114eb0140352e1205933037b5790f271d2a12b10a4aeecaab85a7"} Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.943837 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-5578x" podStartSLOduration=3.4055078610000002 podStartE2EDuration="40.943819705s" podCreationTimestamp="2025-10-04 04:53:00 +0000 UTC" firstStartedPulling="2025-10-04 04:53:02.21134115 +0000 UTC m=+1216.059016618" lastFinishedPulling="2025-10-04 04:53:39.749653004 +0000 UTC m=+1253.597328462" observedRunningTime="2025-10-04 04:53:40.940032624 +0000 UTC m=+1254.787708092" watchObservedRunningTime="2025-10-04 04:53:40.943819705 +0000 UTC m=+1254.791495173" Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.975306 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-55b5954cd6-68rnc" podStartSLOduration=25.90631135 podStartE2EDuration="27.975286761s" podCreationTimestamp="2025-10-04 04:53:13 +0000 UTC" firstStartedPulling="2025-10-04 04:53:37.830636179 +0000 UTC m=+1251.678311647" lastFinishedPulling="2025-10-04 04:53:39.89961159 +0000 UTC m=+1253.747287058" observedRunningTime="2025-10-04 04:53:40.967134634 +0000 UTC m=+1254.814810102" watchObservedRunningTime="2025-10-04 04:53:40.975286761 +0000 UTC m=+1254.822962219" Oct 04 04:53:40 crc kubenswrapper[4992]: I1004 04:53:40.990784 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" podStartSLOduration=35.990767232 podStartE2EDuration="35.990767232s" podCreationTimestamp="2025-10-04 04:53:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:40.988919973 +0000 UTC m=+1254.836595441" watchObservedRunningTime="2025-10-04 04:53:40.990767232 +0000 UTC m=+1254.838442700" Oct 04 04:53:41 crc kubenswrapper[4992]: I1004 04:53:41.956603 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01f98839-df68-4910-9a54-2236f5f1e6e6","Type":"ContainerStarted","Data":"7f1c361bfaed38c2047fd57087b194b3b9e0c35182fd05e15db778c9a4fd8512"} Oct 04 04:53:41 crc kubenswrapper[4992]: I1004 04:53:41.956828 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerName="glance-httpd" containerID="cri-o://7f1c361bfaed38c2047fd57087b194b3b9e0c35182fd05e15db778c9a4fd8512" gracePeriod=30 Oct 04 04:53:41 crc kubenswrapper[4992]: I1004 04:53:41.956748 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerName="glance-log" containerID="cri-o://82da68eb686e1cb14a3ace3813583a92e7028695d961ea913f66b49352c4f75e" gracePeriod=30 Oct 04 04:53:41 crc kubenswrapper[4992]: I1004 04:53:41.961464 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerName="glance-log" containerID="cri-o://bd044f27ab0c748ae2c543e1149f496b4a1ca441f2268d7d4c64bfbd930ceabc" gracePeriod=30 Oct 04 04:53:41 crc kubenswrapper[4992]: I1004 04:53:41.961559 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a","Type":"ContainerStarted","Data":"f680749a7eb3637d19de5a388e2f5c38d50edcf0b950fdfd07e285849960d1d0"} Oct 04 04:53:41 crc kubenswrapper[4992]: I1004 04:53:41.961581 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a","Type":"ContainerStarted","Data":"bd044f27ab0c748ae2c543e1149f496b4a1ca441f2268d7d4c64bfbd930ceabc"} Oct 04 04:53:41 crc kubenswrapper[4992]: I1004 04:53:41.962149 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerName="glance-httpd" containerID="cri-o://f680749a7eb3637d19de5a388e2f5c38d50edcf0b950fdfd07e285849960d1d0" gracePeriod=30 Oct 04 04:53:41 crc kubenswrapper[4992]: I1004 04:53:41.979558 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=35.979536523 podStartE2EDuration="35.979536523s" podCreationTimestamp="2025-10-04 04:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:41.979124083 +0000 UTC m=+1255.826799551" watchObservedRunningTime="2025-10-04 04:53:41.979536523 +0000 UTC m=+1255.827211991" Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.006435 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=37.006417218 podStartE2EDuration="37.006417218s" podCreationTimestamp="2025-10-04 04:53:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:42.003987343 +0000 UTC m=+1255.851662801" watchObservedRunningTime="2025-10-04 04:53:42.006417218 +0000 UTC m=+1255.854092686" Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.970508 4992 generic.go:334] "Generic (PLEG): container finished" podID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerID="7f1c361bfaed38c2047fd57087b194b3b9e0c35182fd05e15db778c9a4fd8512" exitCode=0 Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.970873 4992 generic.go:334] "Generic (PLEG): container finished" podID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerID="82da68eb686e1cb14a3ace3813583a92e7028695d961ea913f66b49352c4f75e" exitCode=143 Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.970710 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01f98839-df68-4910-9a54-2236f5f1e6e6","Type":"ContainerDied","Data":"7f1c361bfaed38c2047fd57087b194b3b9e0c35182fd05e15db778c9a4fd8512"} Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.970975 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01f98839-df68-4910-9a54-2236f5f1e6e6","Type":"ContainerDied","Data":"82da68eb686e1cb14a3ace3813583a92e7028695d961ea913f66b49352c4f75e"} Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.973985 4992 generic.go:334] "Generic (PLEG): container finished" podID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerID="f680749a7eb3637d19de5a388e2f5c38d50edcf0b950fdfd07e285849960d1d0" exitCode=0 Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.974014 4992 generic.go:334] "Generic (PLEG): container finished" podID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerID="bd044f27ab0c748ae2c543e1149f496b4a1ca441f2268d7d4c64bfbd930ceabc" exitCode=143 Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.974033 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a","Type":"ContainerDied","Data":"f680749a7eb3637d19de5a388e2f5c38d50edcf0b950fdfd07e285849960d1d0"} Oct 04 04:53:42 crc kubenswrapper[4992]: I1004 04:53:42.974057 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a","Type":"ContainerDied","Data":"bd044f27ab0c748ae2c543e1149f496b4a1ca441f2268d7d4c64bfbd930ceabc"} Oct 04 04:53:44 crc kubenswrapper[4992]: I1004 04:53:44.056711 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:44 crc kubenswrapper[4992]: I1004 04:53:44.057006 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:53:46 crc kubenswrapper[4992]: I1004 04:53:46.331084 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:53:46 crc kubenswrapper[4992]: I1004 04:53:46.407915 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4jk5d"] Oct 04 04:53:46 crc kubenswrapper[4992]: I1004 04:53:46.408321 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-4jk5d" podUID="365fab3e-1695-4e8e-a362-485daebb7f46" containerName="dnsmasq-dns" containerID="cri-o://615cc9b13c050a32ab4c27c138dd9d554f693391f24b406b3d14e1d3f48e6cd7" gracePeriod=10 Oct 04 04:53:46 crc kubenswrapper[4992]: I1004 04:53:46.769921 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-4jk5d" podUID="365fab3e-1695-4e8e-a362-485daebb7f46" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.111:5353: connect: connection refused" Oct 04 04:53:48 crc kubenswrapper[4992]: I1004 04:53:48.028903 4992 generic.go:334] "Generic (PLEG): container finished" podID="365fab3e-1695-4e8e-a362-485daebb7f46" containerID="615cc9b13c050a32ab4c27c138dd9d554f693391f24b406b3d14e1d3f48e6cd7" exitCode=0 Oct 04 04:53:48 crc kubenswrapper[4992]: I1004 04:53:48.028969 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4jk5d" event={"ID":"365fab3e-1695-4e8e-a362-485daebb7f46","Type":"ContainerDied","Data":"615cc9b13c050a32ab4c27c138dd9d554f693391f24b406b3d14e1d3f48e6cd7"} Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.176995 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.178890 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220246 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-config-data\") pod \"01f98839-df68-4910-9a54-2236f5f1e6e6\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220287 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-config-data\") pod \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220327 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-combined-ca-bundle\") pod \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220374 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-logs\") pod \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220392 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-httpd-run\") pod \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220437 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cdrg\" (UniqueName: \"kubernetes.io/projected/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-kube-api-access-8cdrg\") pod \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220473 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-scripts\") pod \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220501 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9f88\" (UniqueName: \"kubernetes.io/projected/01f98839-df68-4910-9a54-2236f5f1e6e6-kube-api-access-p9f88\") pod \"01f98839-df68-4910-9a54-2236f5f1e6e6\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220525 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-combined-ca-bundle\") pod \"01f98839-df68-4910-9a54-2236f5f1e6e6\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220592 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220614 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"01f98839-df68-4910-9a54-2236f5f1e6e6\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220636 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-scripts\") pod \"01f98839-df68-4910-9a54-2236f5f1e6e6\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220654 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-logs\") pod \"01f98839-df68-4910-9a54-2236f5f1e6e6\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.220732 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-httpd-run\") pod \"01f98839-df68-4910-9a54-2236f5f1e6e6\" (UID: \"01f98839-df68-4910-9a54-2236f5f1e6e6\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.222190 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-logs" (OuterVolumeSpecName: "logs") pod "01f98839-df68-4910-9a54-2236f5f1e6e6" (UID: "01f98839-df68-4910-9a54-2236f5f1e6e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.231962 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" (UID: "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.232478 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-logs" (OuterVolumeSpecName: "logs") pod "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" (UID: "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.233520 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "01f98839-df68-4910-9a54-2236f5f1e6e6" (UID: "01f98839-df68-4910-9a54-2236f5f1e6e6"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.234216 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.234312 4992 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.234931 4992 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.235547 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.236536 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "01f98839-df68-4910-9a54-2236f5f1e6e6" (UID: "01f98839-df68-4910-9a54-2236f5f1e6e6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.239588 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-scripts" (OuterVolumeSpecName: "scripts") pod "01f98839-df68-4910-9a54-2236f5f1e6e6" (UID: "01f98839-df68-4910-9a54-2236f5f1e6e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.245761 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-kube-api-access-8cdrg" (OuterVolumeSpecName: "kube-api-access-8cdrg") pod "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" (UID: "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a"). InnerVolumeSpecName "kube-api-access-8cdrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.254629 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01f98839-df68-4910-9a54-2236f5f1e6e6-kube-api-access-p9f88" (OuterVolumeSpecName: "kube-api-access-p9f88") pod "01f98839-df68-4910-9a54-2236f5f1e6e6" (UID: "01f98839-df68-4910-9a54-2236f5f1e6e6"). InnerVolumeSpecName "kube-api-access-p9f88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.259733 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" (UID: "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.261510 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-scripts" (OuterVolumeSpecName: "scripts") pod "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" (UID: "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.285312 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" (UID: "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.305997 4992 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.306459 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01f98839-df68-4910-9a54-2236f5f1e6e6" (UID: "01f98839-df68-4910-9a54-2236f5f1e6e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.347777 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-config-data" (OuterVolumeSpecName: "config-data") pod "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" (UID: "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348130 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-config-data\") pod \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\" (UID: \"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a\") " Oct 04 04:53:49 crc kubenswrapper[4992]: W1004 04:53:49.348334 4992 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a/volumes/kubernetes.io~secret/config-data Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348400 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-config-data" (OuterVolumeSpecName: "config-data") pod "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" (UID: "dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348654 4992 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/01f98839-df68-4910-9a54-2236f5f1e6e6-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348681 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348692 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348707 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cdrg\" (UniqueName: \"kubernetes.io/projected/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-kube-api-access-8cdrg\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348718 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348730 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9f88\" (UniqueName: \"kubernetes.io/projected/01f98839-df68-4910-9a54-2236f5f1e6e6-kube-api-access-p9f88\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348743 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348778 4992 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348791 4992 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.348802 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.354536 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-config-data" (OuterVolumeSpecName: "config-data") pod "01f98839-df68-4910-9a54-2236f5f1e6e6" (UID: "01f98839-df68-4910-9a54-2236f5f1e6e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.372569 4992 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.441411 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.450530 4992 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.450565 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01f98839-df68-4910-9a54-2236f5f1e6e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.551959 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-dns-svc\") pod \"365fab3e-1695-4e8e-a362-485daebb7f46\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.552349 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-config\") pod \"365fab3e-1695-4e8e-a362-485daebb7f46\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.552395 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-sb\") pod \"365fab3e-1695-4e8e-a362-485daebb7f46\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.552447 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-nb\") pod \"365fab3e-1695-4e8e-a362-485daebb7f46\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.552474 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf9tz\" (UniqueName: \"kubernetes.io/projected/365fab3e-1695-4e8e-a362-485daebb7f46-kube-api-access-vf9tz\") pod \"365fab3e-1695-4e8e-a362-485daebb7f46\" (UID: \"365fab3e-1695-4e8e-a362-485daebb7f46\") " Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.557352 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/365fab3e-1695-4e8e-a362-485daebb7f46-kube-api-access-vf9tz" (OuterVolumeSpecName: "kube-api-access-vf9tz") pod "365fab3e-1695-4e8e-a362-485daebb7f46" (UID: "365fab3e-1695-4e8e-a362-485daebb7f46"). InnerVolumeSpecName "kube-api-access-vf9tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.596811 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "365fab3e-1695-4e8e-a362-485daebb7f46" (UID: "365fab3e-1695-4e8e-a362-485daebb7f46"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.598090 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "365fab3e-1695-4e8e-a362-485daebb7f46" (UID: "365fab3e-1695-4e8e-a362-485daebb7f46"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.601491 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "365fab3e-1695-4e8e-a362-485daebb7f46" (UID: "365fab3e-1695-4e8e-a362-485daebb7f46"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.602047 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-config" (OuterVolumeSpecName: "config") pod "365fab3e-1695-4e8e-a362-485daebb7f46" (UID: "365fab3e-1695-4e8e-a362-485daebb7f46"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.654397 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.654441 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.654452 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.654466 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/365fab3e-1695-4e8e-a362-485daebb7f46-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:49 crc kubenswrapper[4992]: I1004 04:53:49.654478 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf9tz\" (UniqueName: \"kubernetes.io/projected/365fab3e-1695-4e8e-a362-485daebb7f46-kube-api-access-vf9tz\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.074462 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a","Type":"ContainerDied","Data":"9cabb92cf0277db5caf1115033e45f70c82d6c337d20a07f104e86d0991406ca"} Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.074538 4992 scope.go:117] "RemoveContainer" containerID="f680749a7eb3637d19de5a388e2f5c38d50edcf0b950fdfd07e285849960d1d0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.074731 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.082806 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"01f98839-df68-4910-9a54-2236f5f1e6e6","Type":"ContainerDied","Data":"2e8d9e183b446633cf3082f00cc8cd70a0fadaf1f3834b6f771b8241dfad44be"} Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.082954 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.087178 4992 generic.go:334] "Generic (PLEG): container finished" podID="dd8feb2f-147b-43d8-ab87-58703170e959" containerID="debf3cba486bfdfb01da926716ae7e614f2f71ca37382b9cabbcb1fb44e12bb8" exitCode=0 Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.087278 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ln7d7" event={"ID":"dd8feb2f-147b-43d8-ab87-58703170e959","Type":"ContainerDied","Data":"debf3cba486bfdfb01da926716ae7e614f2f71ca37382b9cabbcb1fb44e12bb8"} Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.098426 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652","Type":"ContainerStarted","Data":"d6a75282c1df8dc25bc27d0c7fd5eeea686447968833ec972d161846603c177b"} Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.104625 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-4jk5d" event={"ID":"365fab3e-1695-4e8e-a362-485daebb7f46","Type":"ContainerDied","Data":"d3174bbb8fcfefcba7d9dffaacebca5ad1c03ffad9bda4a7d2f982bc9f4e169e"} Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.104738 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-4jk5d" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.244481 4992 scope.go:117] "RemoveContainer" containerID="bd044f27ab0c748ae2c543e1149f496b4a1ca441f2268d7d4c64bfbd930ceabc" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.247411 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.259976 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.270004 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.289575 4992 scope.go:117] "RemoveContainer" containerID="7f1c361bfaed38c2047fd57087b194b3b9e0c35182fd05e15db778c9a4fd8512" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.295259 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:50 crc kubenswrapper[4992]: E1004 04:53:50.295788 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerName="glance-log" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.295804 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerName="glance-log" Oct 04 04:53:50 crc kubenswrapper[4992]: E1004 04:53:50.295826 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerName="glance-httpd" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.295834 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerName="glance-httpd" Oct 04 04:53:50 crc kubenswrapper[4992]: E1004 04:53:50.295849 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.295857 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" Oct 04 04:53:50 crc kubenswrapper[4992]: E1004 04:53:50.295870 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="365fab3e-1695-4e8e-a362-485daebb7f46" containerName="init" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.295878 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="365fab3e-1695-4e8e-a362-485daebb7f46" containerName="init" Oct 04 04:53:50 crc kubenswrapper[4992]: E1004 04:53:50.295888 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="init" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.295895 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="init" Oct 04 04:53:50 crc kubenswrapper[4992]: E1004 04:53:50.295907 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="365fab3e-1695-4e8e-a362-485daebb7f46" containerName="dnsmasq-dns" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296041 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="365fab3e-1695-4e8e-a362-485daebb7f46" containerName="dnsmasq-dns" Oct 04 04:53:50 crc kubenswrapper[4992]: E1004 04:53:50.296062 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerName="glance-log" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296071 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerName="glance-log" Oct 04 04:53:50 crc kubenswrapper[4992]: E1004 04:53:50.296090 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerName="glance-httpd" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296098 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerName="glance-httpd" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296303 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerName="glance-log" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296315 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerName="glance-log" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296349 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="365fab3e-1695-4e8e-a362-485daebb7f46" containerName="dnsmasq-dns" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296376 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" containerName="glance-httpd" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296385 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" containerName="glance-httpd" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.296400 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="047d386f-a54b-4cd6-a3a5-764f65f80624" containerName="dnsmasq-dns" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.297459 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.308554 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zjhxk" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.310074 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.310784 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.311003 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.313027 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.346055 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01f98839-df68-4910-9a54-2236f5f1e6e6" path="/var/lib/kubelet/pods/01f98839-df68-4910-9a54-2236f5f1e6e6/volumes" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.346960 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a" path="/var/lib/kubelet/pods/dfe46e73-6ba1-4e3f-95cd-c0ca10848c2a/volumes" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.347750 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4jk5d"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.348094 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.356475 4992 scope.go:117] "RemoveContainer" containerID="82da68eb686e1cb14a3ace3813583a92e7028695d961ea913f66b49352c4f75e" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.368796 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-4jk5d"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.377571 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-logs\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.377738 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.377816 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.377931 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.378024 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.378145 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.378314 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.378550 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llv8m\" (UniqueName: \"kubernetes.io/projected/3539e12b-5589-449d-9600-02d7aab34e88-kube-api-access-llv8m\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.394131 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.396298 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.396527 4992 scope.go:117] "RemoveContainer" containerID="615cc9b13c050a32ab4c27c138dd9d554f693391f24b406b3d14e1d3f48e6cd7" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.398938 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.400531 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.405588 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.431586 4992 scope.go:117] "RemoveContainer" containerID="17e420e4d0d0c301a5c9c64116705a4065e33c585ae24d9d431da21b1784953e" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.480312 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.480444 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.480656 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.480718 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481435 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481494 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481542 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481580 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481629 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481671 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-logs\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481728 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llv8m\" (UniqueName: \"kubernetes.io/projected/3539e12b-5589-449d-9600-02d7aab34e88-kube-api-access-llv8m\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481771 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-logs\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481799 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481855 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.481863 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.484076 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.482596 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-logs\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.484420 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgt7k\" (UniqueName: \"kubernetes.io/projected/9f3378de-a1d7-4af4-8e08-de77091ed6c4-kube-api-access-dgt7k\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.485651 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.486742 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.491468 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.495929 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.501089 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.519889 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llv8m\" (UniqueName: \"kubernetes.io/projected/3539e12b-5589-449d-9600-02d7aab34e88-kube-api-access-llv8m\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.524972 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.586576 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.586675 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.586717 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.586742 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.586813 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-logs\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.586895 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.587027 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.587052 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgt7k\" (UniqueName: \"kubernetes.io/projected/9f3378de-a1d7-4af4-8e08-de77091ed6c4-kube-api-access-dgt7k\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.589011 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-logs\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.589131 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.589638 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.590239 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-config-data\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.592297 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.593952 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.601197 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-scripts\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.610637 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgt7k\" (UniqueName: \"kubernetes.io/projected/9f3378de-a1d7-4af4-8e08-de77091ed6c4-kube-api-access-dgt7k\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.626680 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.647819 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.723600 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.894093 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:53:50 crc kubenswrapper[4992]: I1004 04:53:50.894167 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.117424 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zcz9v" event={"ID":"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6","Type":"ContainerStarted","Data":"00cff6f979663a239fcdc719ccbebebc14bbf88001e8f681e8cb7dfdab02ad74"} Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.192803 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-zcz9v" podStartSLOduration=3.162957271 podStartE2EDuration="51.192783784s" podCreationTimestamp="2025-10-04 04:53:00 +0000 UTC" firstStartedPulling="2025-10-04 04:53:02.037505737 +0000 UTC m=+1215.885181205" lastFinishedPulling="2025-10-04 04:53:50.06733225 +0000 UTC m=+1263.915007718" observedRunningTime="2025-10-04 04:53:51.139521119 +0000 UTC m=+1264.987196607" watchObservedRunningTime="2025-10-04 04:53:51.192783784 +0000 UTC m=+1265.040459252" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.200067 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.320527 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:53:51 crc kubenswrapper[4992]: W1004 04:53:51.372840 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f3378de_a1d7_4af4_8e08_de77091ed6c4.slice/crio-f36d6133b9dac2fcc7a245397e99600c10fc7fbba564f2183a8071653bc80b78 WatchSource:0}: Error finding container f36d6133b9dac2fcc7a245397e99600c10fc7fbba564f2183a8071653bc80b78: Status 404 returned error can't find the container with id f36d6133b9dac2fcc7a245397e99600c10fc7fbba564f2183a8071653bc80b78 Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.534744 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.624427 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-config-data\") pod \"dd8feb2f-147b-43d8-ab87-58703170e959\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.624572 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-fernet-keys\") pod \"dd8feb2f-147b-43d8-ab87-58703170e959\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.624622 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7f9j\" (UniqueName: \"kubernetes.io/projected/dd8feb2f-147b-43d8-ab87-58703170e959-kube-api-access-t7f9j\") pod \"dd8feb2f-147b-43d8-ab87-58703170e959\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.624658 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-combined-ca-bundle\") pod \"dd8feb2f-147b-43d8-ab87-58703170e959\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.624765 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-scripts\") pod \"dd8feb2f-147b-43d8-ab87-58703170e959\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.624789 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-credential-keys\") pod \"dd8feb2f-147b-43d8-ab87-58703170e959\" (UID: \"dd8feb2f-147b-43d8-ab87-58703170e959\") " Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.632450 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "dd8feb2f-147b-43d8-ab87-58703170e959" (UID: "dd8feb2f-147b-43d8-ab87-58703170e959"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.632892 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-scripts" (OuterVolumeSpecName: "scripts") pod "dd8feb2f-147b-43d8-ab87-58703170e959" (UID: "dd8feb2f-147b-43d8-ab87-58703170e959"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.634312 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "dd8feb2f-147b-43d8-ab87-58703170e959" (UID: "dd8feb2f-147b-43d8-ab87-58703170e959"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.634558 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd8feb2f-147b-43d8-ab87-58703170e959-kube-api-access-t7f9j" (OuterVolumeSpecName: "kube-api-access-t7f9j") pod "dd8feb2f-147b-43d8-ab87-58703170e959" (UID: "dd8feb2f-147b-43d8-ab87-58703170e959"). InnerVolumeSpecName "kube-api-access-t7f9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.653775 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd8feb2f-147b-43d8-ab87-58703170e959" (UID: "dd8feb2f-147b-43d8-ab87-58703170e959"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.663996 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-config-data" (OuterVolumeSpecName: "config-data") pod "dd8feb2f-147b-43d8-ab87-58703170e959" (UID: "dd8feb2f-147b-43d8-ab87-58703170e959"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.727303 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.727566 4992 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.727583 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7f9j\" (UniqueName: \"kubernetes.io/projected/dd8feb2f-147b-43d8-ab87-58703170e959-kube-api-access-t7f9j\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.727599 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.727610 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:51 crc kubenswrapper[4992]: I1004 04:53:51.727623 4992 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dd8feb2f-147b-43d8-ab87-58703170e959-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.148791 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ln7d7" event={"ID":"dd8feb2f-147b-43d8-ab87-58703170e959","Type":"ContainerDied","Data":"f01ebf0845ae4d1e15838fe6610586e68ef7291102974cae4ede345e451df768"} Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.149133 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f01ebf0845ae4d1e15838fe6610586e68ef7291102974cae4ede345e451df768" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.149217 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ln7d7" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.153075 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3539e12b-5589-449d-9600-02d7aab34e88","Type":"ContainerStarted","Data":"b92e82a510849611c9aa4065a21358500e150e0b4e146d2526ecfe2f61ae942b"} Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.153812 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3539e12b-5589-449d-9600-02d7aab34e88","Type":"ContainerStarted","Data":"fecf8a58f9bff432b5b6c40dbe3ebcf743a736d9f4a7f43d09d939897099a2c8"} Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.158162 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6dc74865-2rgnq" event={"ID":"94c198d3-0808-4077-951a-fe14a94409ea","Type":"ContainerStarted","Data":"0e838cf84f8b2286de7fee1771b85b59290fdf40b14d9269125d5379345a8111"} Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.158209 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6dc74865-2rgnq" event={"ID":"94c198d3-0808-4077-951a-fe14a94409ea","Type":"ContainerStarted","Data":"5360ba4db45400a267ecaceaa1c0924dabeb2f6d586b197d48d547360639404f"} Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.163098 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f3378de-a1d7-4af4-8e08-de77091ed6c4","Type":"ContainerStarted","Data":"f36d6133b9dac2fcc7a245397e99600c10fc7fbba564f2183a8071653bc80b78"} Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.193783 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c6dc74865-2rgnq" podStartSLOduration=-9223371987.66102 podStartE2EDuration="49.19375563s" podCreationTimestamp="2025-10-04 04:53:03 +0000 UTC" firstStartedPulling="2025-10-04 04:53:04.334027944 +0000 UTC m=+1218.181703412" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:52.17495677 +0000 UTC m=+1266.022632248" watchObservedRunningTime="2025-10-04 04:53:52.19375563 +0000 UTC m=+1266.041431098" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.252783 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5bb778b4c-snbm9"] Oct 04 04:53:52 crc kubenswrapper[4992]: E1004 04:53:52.253347 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd8feb2f-147b-43d8-ab87-58703170e959" containerName="keystone-bootstrap" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.253381 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd8feb2f-147b-43d8-ab87-58703170e959" containerName="keystone-bootstrap" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.253633 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd8feb2f-147b-43d8-ab87-58703170e959" containerName="keystone-bootstrap" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.255458 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.261612 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.261891 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.262031 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.262142 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.262250 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5nrbs" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.262420 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.276163 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5bb778b4c-snbm9"] Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.340232 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-public-tls-certs\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.340317 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-credential-keys\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.340347 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-combined-ca-bundle\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.340441 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvc54\" (UniqueName: \"kubernetes.io/projected/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-kube-api-access-wvc54\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.340477 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-internal-tls-certs\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.340513 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-config-data\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.340540 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-scripts\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.340683 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-fernet-keys\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.344803 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="365fab3e-1695-4e8e-a362-485daebb7f46" path="/var/lib/kubelet/pods/365fab3e-1695-4e8e-a362-485daebb7f46/volumes" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.442163 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-fernet-keys\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.442243 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-public-tls-certs\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.442305 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-credential-keys\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.442327 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-combined-ca-bundle\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.442421 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvc54\" (UniqueName: \"kubernetes.io/projected/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-kube-api-access-wvc54\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.443094 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-internal-tls-certs\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.443162 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-config-data\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.443183 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-scripts\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.447471 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-scripts\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.448981 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-fernet-keys\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.449323 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-public-tls-certs\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.452834 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-credential-keys\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.453110 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-internal-tls-certs\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.455474 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-config-data\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.467800 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-combined-ca-bundle\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.468050 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvc54\" (UniqueName: \"kubernetes.io/projected/e15d6dcc-af83-40dd-b5a2-936c8e5154b3-kube-api-access-wvc54\") pod \"keystone-5bb778b4c-snbm9\" (UID: \"e15d6dcc-af83-40dd-b5a2-936c8e5154b3\") " pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:52 crc kubenswrapper[4992]: I1004 04:53:52.600076 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:53 crc kubenswrapper[4992]: I1004 04:53:53.059198 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5bb778b4c-snbm9"] Oct 04 04:53:53 crc kubenswrapper[4992]: W1004 04:53:53.071064 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode15d6dcc_af83_40dd_b5a2_936c8e5154b3.slice/crio-2cd666038c63261be3ab4fd2953a85c6133023da85f2bf8bc430a34a49936700 WatchSource:0}: Error finding container 2cd666038c63261be3ab4fd2953a85c6133023da85f2bf8bc430a34a49936700: Status 404 returned error can't find the container with id 2cd666038c63261be3ab4fd2953a85c6133023da85f2bf8bc430a34a49936700 Oct 04 04:53:53 crc kubenswrapper[4992]: I1004 04:53:53.180083 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3539e12b-5589-449d-9600-02d7aab34e88","Type":"ContainerStarted","Data":"1623e13998770c5dceaaa0253008e74fad9abf73c5570330dba759a257ad56da"} Oct 04 04:53:53 crc kubenswrapper[4992]: I1004 04:53:53.183666 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f3378de-a1d7-4af4-8e08-de77091ed6c4","Type":"ContainerStarted","Data":"06d326cfffb4e6048ef38f1a2b0139d587f975f8800c25fafe4f106304f25ab1"} Oct 04 04:53:53 crc kubenswrapper[4992]: I1004 04:53:53.185352 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5bb778b4c-snbm9" event={"ID":"e15d6dcc-af83-40dd-b5a2-936c8e5154b3","Type":"ContainerStarted","Data":"2cd666038c63261be3ab4fd2953a85c6133023da85f2bf8bc430a34a49936700"} Oct 04 04:53:53 crc kubenswrapper[4992]: I1004 04:53:53.210891 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.210871584 podStartE2EDuration="3.210871584s" podCreationTimestamp="2025-10-04 04:53:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:53.204000451 +0000 UTC m=+1267.051675939" watchObservedRunningTime="2025-10-04 04:53:53.210871584 +0000 UTC m=+1267.058547052" Oct 04 04:53:53 crc kubenswrapper[4992]: I1004 04:53:53.752129 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:53 crc kubenswrapper[4992]: I1004 04:53:53.757738 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:53:54 crc kubenswrapper[4992]: I1004 04:53:54.058402 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-55b5954cd6-68rnc" podUID="d75279dd-898a-4ffe-85e6-804f066de711" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 04:53:54 crc kubenswrapper[4992]: I1004 04:53:54.198961 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f3378de-a1d7-4af4-8e08-de77091ed6c4","Type":"ContainerStarted","Data":"43c80d655db6f4d7323bea992428e7fe03178268abc9899c78ec35cf517536d6"} Oct 04 04:53:54 crc kubenswrapper[4992]: I1004 04:53:54.201771 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5bb778b4c-snbm9" event={"ID":"e15d6dcc-af83-40dd-b5a2-936c8e5154b3","Type":"ContainerStarted","Data":"4d35fe950287cdd0daabb471f1c1df44f352c29155e69932979cf19a8fa23323"} Oct 04 04:53:54 crc kubenswrapper[4992]: I1004 04:53:54.202077 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:53:54 crc kubenswrapper[4992]: I1004 04:53:54.224139 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.224112554 podStartE2EDuration="4.224112554s" podCreationTimestamp="2025-10-04 04:53:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:54.219460871 +0000 UTC m=+1268.067136339" watchObservedRunningTime="2025-10-04 04:53:54.224112554 +0000 UTC m=+1268.071788022" Oct 04 04:53:54 crc kubenswrapper[4992]: I1004 04:53:54.244815 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5bb778b4c-snbm9" podStartSLOduration=2.244788084 podStartE2EDuration="2.244788084s" podCreationTimestamp="2025-10-04 04:53:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:53:54.238410035 +0000 UTC m=+1268.086085523" watchObservedRunningTime="2025-10-04 04:53:54.244788084 +0000 UTC m=+1268.092463552" Oct 04 04:53:56 crc kubenswrapper[4992]: I1004 04:53:56.223749 4992 generic.go:334] "Generic (PLEG): container finished" podID="15d2bb9c-fea2-48e2-aac2-3f07c2813b25" containerID="a708ae19fdcb6154dfbe928d522052f3b71af3f67a4382c3650b02f3c29e2a7c" exitCode=0 Oct 04 04:53:56 crc kubenswrapper[4992]: I1004 04:53:56.223869 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-srrcj" event={"ID":"15d2bb9c-fea2-48e2-aac2-3f07c2813b25","Type":"ContainerDied","Data":"a708ae19fdcb6154dfbe928d522052f3b71af3f67a4382c3650b02f3c29e2a7c"} Oct 04 04:53:56 crc kubenswrapper[4992]: I1004 04:53:56.227586 4992 generic.go:334] "Generic (PLEG): container finished" podID="d2f4ea76-03f4-49e2-99d8-af57dabb6372" containerID="37d9da7f7b13e8017ebf021988b1d3a87fffa72d2f31f182996f8348347bd403" exitCode=0 Oct 04 04:53:56 crc kubenswrapper[4992]: I1004 04:53:56.227627 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5578x" event={"ID":"d2f4ea76-03f4-49e2-99d8-af57dabb6372","Type":"ContainerDied","Data":"37d9da7f7b13e8017ebf021988b1d3a87fffa72d2f31f182996f8348347bd403"} Oct 04 04:54:00 crc kubenswrapper[4992]: I1004 04:54:00.648310 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:54:00 crc kubenswrapper[4992]: I1004 04:54:00.648696 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:54:00 crc kubenswrapper[4992]: I1004 04:54:00.683167 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:54:00 crc kubenswrapper[4992]: I1004 04:54:00.692744 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:54:00 crc kubenswrapper[4992]: I1004 04:54:00.723811 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:54:00 crc kubenswrapper[4992]: I1004 04:54:00.723852 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:54:00 crc kubenswrapper[4992]: I1004 04:54:00.754509 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:54:00 crc kubenswrapper[4992]: I1004 04:54:00.793637 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.272762 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.272907 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.272921 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.272930 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.564511 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5578x" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.618400 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-combined-ca-bundle\") pod \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.618603 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-scripts\") pod \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.618639 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-config-data\") pod \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.618689 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw87m\" (UniqueName: \"kubernetes.io/projected/d2f4ea76-03f4-49e2-99d8-af57dabb6372-kube-api-access-bw87m\") pod \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.618753 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2f4ea76-03f4-49e2-99d8-af57dabb6372-logs\") pod \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\" (UID: \"d2f4ea76-03f4-49e2-99d8-af57dabb6372\") " Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.619772 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2f4ea76-03f4-49e2-99d8-af57dabb6372-logs" (OuterVolumeSpecName: "logs") pod "d2f4ea76-03f4-49e2-99d8-af57dabb6372" (UID: "d2f4ea76-03f4-49e2-99d8-af57dabb6372"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.624171 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f4ea76-03f4-49e2-99d8-af57dabb6372-kube-api-access-bw87m" (OuterVolumeSpecName: "kube-api-access-bw87m") pod "d2f4ea76-03f4-49e2-99d8-af57dabb6372" (UID: "d2f4ea76-03f4-49e2-99d8-af57dabb6372"). InnerVolumeSpecName "kube-api-access-bw87m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.641541 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-scripts" (OuterVolumeSpecName: "scripts") pod "d2f4ea76-03f4-49e2-99d8-af57dabb6372" (UID: "d2f4ea76-03f4-49e2-99d8-af57dabb6372"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.666157 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-config-data" (OuterVolumeSpecName: "config-data") pod "d2f4ea76-03f4-49e2-99d8-af57dabb6372" (UID: "d2f4ea76-03f4-49e2-99d8-af57dabb6372"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.714010 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2f4ea76-03f4-49e2-99d8-af57dabb6372" (UID: "d2f4ea76-03f4-49e2-99d8-af57dabb6372"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.725256 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.725291 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.725301 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw87m\" (UniqueName: \"kubernetes.io/projected/d2f4ea76-03f4-49e2-99d8-af57dabb6372-kube-api-access-bw87m\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.725315 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2f4ea76-03f4-49e2-99d8-af57dabb6372-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:01 crc kubenswrapper[4992]: I1004 04:54:01.725324 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f4ea76-03f4-49e2-99d8-af57dabb6372-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.221239 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-srrcj" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.300875 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-srrcj" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.300854 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-srrcj" event={"ID":"15d2bb9c-fea2-48e2-aac2-3f07c2813b25","Type":"ContainerDied","Data":"aeb93c0d887d6a84f3128995b7c0a4b5e6e555fd8e148dd83f0d0061e70b21bc"} Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.301238 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aeb93c0d887d6a84f3128995b7c0a4b5e6e555fd8e148dd83f0d0061e70b21bc" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.310885 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5578x" event={"ID":"d2f4ea76-03f4-49e2-99d8-af57dabb6372","Type":"ContainerDied","Data":"776eef29845ab66c45342c8f52026f7c08373b9e046b88a644833112f4817939"} Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.310926 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="776eef29845ab66c45342c8f52026f7c08373b9e046b88a644833112f4817939" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.311162 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5578x" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.337164 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hff8g\" (UniqueName: \"kubernetes.io/projected/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-kube-api-access-hff8g\") pod \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.337242 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-combined-ca-bundle\") pod \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.337310 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-db-sync-config-data\") pod \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\" (UID: \"15d2bb9c-fea2-48e2-aac2-3f07c2813b25\") " Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.353186 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-kube-api-access-hff8g" (OuterVolumeSpecName: "kube-api-access-hff8g") pod "15d2bb9c-fea2-48e2-aac2-3f07c2813b25" (UID: "15d2bb9c-fea2-48e2-aac2-3f07c2813b25"). InnerVolumeSpecName "kube-api-access-hff8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.369679 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "15d2bb9c-fea2-48e2-aac2-3f07c2813b25" (UID: "15d2bb9c-fea2-48e2-aac2-3f07c2813b25"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.373990 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15d2bb9c-fea2-48e2-aac2-3f07c2813b25" (UID: "15d2bb9c-fea2-48e2-aac2-3f07c2813b25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.442402 4992 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.442788 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hff8g\" (UniqueName: \"kubernetes.io/projected/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-kube-api-access-hff8g\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.442827 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d2bb9c-fea2-48e2-aac2-3f07c2813b25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:02 crc kubenswrapper[4992]: E1004 04:54:02.580789 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.680135 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-689c5fd8b4-p4lxm"] Oct 04 04:54:02 crc kubenswrapper[4992]: E1004 04:54:02.680801 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f4ea76-03f4-49e2-99d8-af57dabb6372" containerName="placement-db-sync" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.680826 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f4ea76-03f4-49e2-99d8-af57dabb6372" containerName="placement-db-sync" Oct 04 04:54:02 crc kubenswrapper[4992]: E1004 04:54:02.680846 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d2bb9c-fea2-48e2-aac2-3f07c2813b25" containerName="barbican-db-sync" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.680855 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d2bb9c-fea2-48e2-aac2-3f07c2813b25" containerName="barbican-db-sync" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.681096 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="15d2bb9c-fea2-48e2-aac2-3f07c2813b25" containerName="barbican-db-sync" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.681128 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2f4ea76-03f4-49e2-99d8-af57dabb6372" containerName="placement-db-sync" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.682255 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.685877 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.686153 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-cb2c8" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.686396 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.686512 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.686635 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.695637 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-689c5fd8b4-p4lxm"] Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.748710 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-internal-tls-certs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.748832 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185b6e92-a7f2-4821-aed0-c732a0dc0e65-logs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.748869 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-scripts\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.748979 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-public-tls-certs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.749012 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-config-data\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.749037 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgfqr\" (UniqueName: \"kubernetes.io/projected/185b6e92-a7f2-4821-aed0-c732a0dc0e65-kube-api-access-vgfqr\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.749059 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-combined-ca-bundle\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.851013 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-internal-tls-certs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.851136 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185b6e92-a7f2-4821-aed0-c732a0dc0e65-logs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.851172 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-scripts\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.851201 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-public-tls-certs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.851234 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-config-data\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.851264 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgfqr\" (UniqueName: \"kubernetes.io/projected/185b6e92-a7f2-4821-aed0-c732a0dc0e65-kube-api-access-vgfqr\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.851290 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-combined-ca-bundle\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.852521 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185b6e92-a7f2-4821-aed0-c732a0dc0e65-logs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.855418 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-internal-tls-certs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.855817 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-scripts\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.859840 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-public-tls-certs\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.861146 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-config-data\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.862139 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185b6e92-a7f2-4821-aed0-c732a0dc0e65-combined-ca-bundle\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:02 crc kubenswrapper[4992]: I1004 04:54:02.879959 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgfqr\" (UniqueName: \"kubernetes.io/projected/185b6e92-a7f2-4821-aed0-c732a0dc0e65-kube-api-access-vgfqr\") pod \"placement-689c5fd8b4-p4lxm\" (UID: \"185b6e92-a7f2-4821-aed0-c732a0dc0e65\") " pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.005187 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.330738 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652","Type":"ContainerStarted","Data":"9fa907e890824e61aa2637b55000791843dfa8d47b50f4589871014f0b17f79d"} Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.331119 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.330981 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="proxy-httpd" containerID="cri-o://9fa907e890824e61aa2637b55000791843dfa8d47b50f4589871014f0b17f79d" gracePeriod=30 Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.330916 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="ceilometer-notification-agent" containerID="cri-o://301352d0463114eb0140352e1205933037b5790f271d2a12b10a4aeecaab85a7" gracePeriod=30 Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.330982 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="sg-core" containerID="cri-o://d6a75282c1df8dc25bc27d0c7fd5eeea686447968833ec972d161846603c177b" gracePeriod=30 Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.531420 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5ff4b555c4-vn5np"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.533912 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.553531 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.553785 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.554018 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fvv86" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.564448 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-689c5fd8b4-p4lxm"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.568053 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-config-data\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.568187 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-config-data-custom\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.568279 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz5pj\" (UniqueName: \"kubernetes.io/projected/a4b0448a-9b84-41c8-a87f-fbc17f92411b-kube-api-access-pz5pj\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.568433 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4b0448a-9b84-41c8-a87f-fbc17f92411b-logs\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.568496 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-combined-ca-bundle\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.579501 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5ff4b555c4-vn5np"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.586726 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-775b66879-7sgzw"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.588629 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.590480 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.609091 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-775b66879-7sgzw"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671466 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhp56\" (UniqueName: \"kubernetes.io/projected/70af60aa-3bee-44bf-874e-648aaefe032d-kube-api-access-bhp56\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671544 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4b0448a-9b84-41c8-a87f-fbc17f92411b-logs\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671588 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-combined-ca-bundle\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671623 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-combined-ca-bundle\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671710 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-config-data\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671741 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70af60aa-3bee-44bf-874e-648aaefe032d-logs\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671779 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-config-data\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671818 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-config-data-custom\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671868 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz5pj\" (UniqueName: \"kubernetes.io/projected/a4b0448a-9b84-41c8-a87f-fbc17f92411b-kube-api-access-pz5pj\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.671900 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-config-data-custom\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.675328 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4b0448a-9b84-41c8-a87f-fbc17f92411b-logs\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.696024 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-config-data-custom\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.699247 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-combined-ca-bundle\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.722585 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4b0448a-9b84-41c8-a87f-fbc17f92411b-config-data\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.749439 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-dhd5n"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.752249 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.758824 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz5pj\" (UniqueName: \"kubernetes.io/projected/a4b0448a-9b84-41c8-a87f-fbc17f92411b-kube-api-access-pz5pj\") pod \"barbican-keystone-listener-5ff4b555c4-vn5np\" (UID: \"a4b0448a-9b84-41c8-a87f-fbc17f92411b\") " pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774489 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-combined-ca-bundle\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774597 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774653 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70af60aa-3bee-44bf-874e-648aaefe032d-logs\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774680 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774717 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-config-data\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774779 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-config\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774825 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-config-data-custom\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774887 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774920 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhp56\" (UniqueName: \"kubernetes.io/projected/70af60aa-3bee-44bf-874e-648aaefe032d-kube-api-access-bhp56\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774950 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxzml\" (UniqueName: \"kubernetes.io/projected/8e07e645-fbe6-461a-96a2-027d28c6c993-kube-api-access-lxzml\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.774981 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.777295 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c6dc74865-2rgnq" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon" probeResult="failure" output="Get \"http://10.217.0.143:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8080: connect: connection refused" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.778289 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.778597 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-dhd5n"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.801780 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70af60aa-3bee-44bf-874e-648aaefe032d-logs\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.819198 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-config-data-custom\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.819411 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-combined-ca-bundle\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.820172 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70af60aa-3bee-44bf-874e-648aaefe032d-config-data\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.831330 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhp56\" (UniqueName: \"kubernetes.io/projected/70af60aa-3bee-44bf-874e-648aaefe032d-kube-api-access-bhp56\") pod \"barbican-worker-775b66879-7sgzw\" (UID: \"70af60aa-3bee-44bf-874e-648aaefe032d\") " pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.847950 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-bcc7d5cbb-phnsf"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.855035 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-bcc7d5cbb-phnsf"] Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.855147 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.864106 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.877751 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.877807 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxzml\" (UniqueName: \"kubernetes.io/projected/8e07e645-fbe6-461a-96a2-027d28c6c993-kube-api-access-lxzml\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.877843 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.877925 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-combined-ca-bundle\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.877959 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data-custom\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.878093 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.878120 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.878243 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79rzn\" (UniqueName: \"kubernetes.io/projected/57dc2121-ed97-4dea-903f-6d3b92b539ff-kube-api-access-79rzn\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.878414 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.878532 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-config\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.879033 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dc2121-ed97-4dea-903f-6d3b92b539ff-logs\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.880498 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.882876 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-config\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.889511 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.890700 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.896059 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.901884 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxzml\" (UniqueName: \"kubernetes.io/projected/8e07e645-fbe6-461a-96a2-027d28c6c993-kube-api-access-lxzml\") pod \"dnsmasq-dns-59d5ff467f-dhd5n\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.981771 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.981829 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79rzn\" (UniqueName: \"kubernetes.io/projected/57dc2121-ed97-4dea-903f-6d3b92b539ff-kube-api-access-79rzn\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.981931 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dc2121-ed97-4dea-903f-6d3b92b539ff-logs\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.982029 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-combined-ca-bundle\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.982077 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data-custom\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.983074 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dc2121-ed97-4dea-903f-6d3b92b539ff-logs\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:03 crc kubenswrapper[4992]: I1004 04:54:03.989060 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-combined-ca-bundle\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.009014 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data-custom\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.009408 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79rzn\" (UniqueName: \"kubernetes.io/projected/57dc2121-ed97-4dea-903f-6d3b92b539ff-kube-api-access-79rzn\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.009549 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data\") pod \"barbican-api-bcc7d5cbb-phnsf\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.132772 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-775b66879-7sgzw" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.148012 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.160082 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.160196 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.169760 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.218948 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.362767 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5ff4b555c4-vn5np"] Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.377280 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.377445 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.379671 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:54:04 crc kubenswrapper[4992]: W1004 04:54:04.424498 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4b0448a_9b84_41c8_a87f_fbc17f92411b.slice/crio-2168ecaee6dc53f6f3d34a91a3ac8b88bf84088e47e513ee48f6616abb39874f WatchSource:0}: Error finding container 2168ecaee6dc53f6f3d34a91a3ac8b88bf84088e47e513ee48f6616abb39874f: Status 404 returned error can't find the container with id 2168ecaee6dc53f6f3d34a91a3ac8b88bf84088e47e513ee48f6616abb39874f Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.431970 4992 generic.go:334] "Generic (PLEG): container finished" podID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerID="9fa907e890824e61aa2637b55000791843dfa8d47b50f4589871014f0b17f79d" exitCode=0 Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.432011 4992 generic.go:334] "Generic (PLEG): container finished" podID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerID="d6a75282c1df8dc25bc27d0c7fd5eeea686447968833ec972d161846603c177b" exitCode=2 Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.432018 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652","Type":"ContainerDied","Data":"9fa907e890824e61aa2637b55000791843dfa8d47b50f4589871014f0b17f79d"} Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.432072 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652","Type":"ContainerDied","Data":"d6a75282c1df8dc25bc27d0c7fd5eeea686447968833ec972d161846603c177b"} Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.475060 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-689c5fd8b4-p4lxm" event={"ID":"185b6e92-a7f2-4821-aed0-c732a0dc0e65","Type":"ContainerStarted","Data":"3894bb079c7a932541ac7a3dad1e05f5f5c994da9f663806efb031e95e697329"} Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.475114 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-689c5fd8b4-p4lxm" event={"ID":"185b6e92-a7f2-4821-aed0-c732a0dc0e65","Type":"ContainerStarted","Data":"43b8ca96d9d44cc77348eae901f7024bfcc1155aac0e14083fbbab3629fdffb7"} Oct 04 04:54:04 crc kubenswrapper[4992]: I1004 04:54:04.935949 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-dhd5n"] Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.192163 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-775b66879-7sgzw"] Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.288689 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-bcc7d5cbb-phnsf"] Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.488525 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" event={"ID":"a4b0448a-9b84-41c8-a87f-fbc17f92411b","Type":"ContainerStarted","Data":"2168ecaee6dc53f6f3d34a91a3ac8b88bf84088e47e513ee48f6616abb39874f"} Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.494876 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-775b66879-7sgzw" event={"ID":"70af60aa-3bee-44bf-874e-648aaefe032d","Type":"ContainerStarted","Data":"b834f875f2ffe77c170d746fab77108d843a091066cf9f30be37180199f2e92b"} Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.504942 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-689c5fd8b4-p4lxm" event={"ID":"185b6e92-a7f2-4821-aed0-c732a0dc0e65","Type":"ContainerStarted","Data":"6097a2a52ee473c9d664cea6c0635e5adde0a08248fce207bdc34a43892e4542"} Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.505131 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.505321 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.508548 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcc7d5cbb-phnsf" event={"ID":"57dc2121-ed97-4dea-903f-6d3b92b539ff","Type":"ContainerStarted","Data":"3aab48c089737e05c712dfde2e52b1d3f10d72ff94a50dc4108fd05c4a25156a"} Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.515739 4992 generic.go:334] "Generic (PLEG): container finished" podID="8e07e645-fbe6-461a-96a2-027d28c6c993" containerID="70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746" exitCode=0 Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.515839 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" event={"ID":"8e07e645-fbe6-461a-96a2-027d28c6c993","Type":"ContainerDied","Data":"70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746"} Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.515893 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" event={"ID":"8e07e645-fbe6-461a-96a2-027d28c6c993","Type":"ContainerStarted","Data":"ff53c5d5d7247afc72cc11977dcad12a18ca9fc3db12cffe9343e93135700b75"} Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.523823 4992 generic.go:334] "Generic (PLEG): container finished" podID="0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" containerID="00cff6f979663a239fcdc719ccbebebc14bbf88001e8f681e8cb7dfdab02ad74" exitCode=0 Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.524751 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zcz9v" event={"ID":"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6","Type":"ContainerDied","Data":"00cff6f979663a239fcdc719ccbebebc14bbf88001e8f681e8cb7dfdab02ad74"} Oct 04 04:54:05 crc kubenswrapper[4992]: I1004 04:54:05.551537 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-689c5fd8b4-p4lxm" podStartSLOduration=3.551514238 podStartE2EDuration="3.551514238s" podCreationTimestamp="2025-10-04 04:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:05.535373309 +0000 UTC m=+1279.383048787" watchObservedRunningTime="2025-10-04 04:54:05.551514238 +0000 UTC m=+1279.399189706" Oct 04 04:54:06 crc kubenswrapper[4992]: I1004 04:54:06.550040 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcc7d5cbb-phnsf" event={"ID":"57dc2121-ed97-4dea-903f-6d3b92b539ff","Type":"ContainerStarted","Data":"d3a0eb567bc650c5d526e462af54066bd733518d0f0d841e5eaf06d0521a1fed"} Oct 04 04:54:06 crc kubenswrapper[4992]: I1004 04:54:06.550394 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcc7d5cbb-phnsf" event={"ID":"57dc2121-ed97-4dea-903f-6d3b92b539ff","Type":"ContainerStarted","Data":"702f1d81b231f96c38514f0e51d7f53cdd3cafbd3878d84a9a298ed152f1ef87"} Oct 04 04:54:06 crc kubenswrapper[4992]: I1004 04:54:06.550734 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:06 crc kubenswrapper[4992]: I1004 04:54:06.550765 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:06 crc kubenswrapper[4992]: I1004 04:54:06.570545 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" event={"ID":"8e07e645-fbe6-461a-96a2-027d28c6c993","Type":"ContainerStarted","Data":"e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87"} Oct 04 04:54:06 crc kubenswrapper[4992]: I1004 04:54:06.570828 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:06 crc kubenswrapper[4992]: I1004 04:54:06.585436 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-bcc7d5cbb-phnsf" podStartSLOduration=3.585409469 podStartE2EDuration="3.585409469s" podCreationTimestamp="2025-10-04 04:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:06.57304567 +0000 UTC m=+1280.420721138" watchObservedRunningTime="2025-10-04 04:54:06.585409469 +0000 UTC m=+1280.433084947" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.349095 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.367425 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" podStartSLOduration=4.367352292 podStartE2EDuration="4.367352292s" podCreationTimestamp="2025-10-04 04:54:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:06.622862894 +0000 UTC m=+1280.470538362" watchObservedRunningTime="2025-10-04 04:54:07.367352292 +0000 UTC m=+1281.215027760" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.414791 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-etc-machine-id\") pod \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.414915 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-db-sync-config-data\") pod \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.414966 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-scripts\") pod \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.415025 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-config-data\") pod \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.415067 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpwvw\" (UniqueName: \"kubernetes.io/projected/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-kube-api-access-wpwvw\") pod \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.415162 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-combined-ca-bundle\") pod \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\" (UID: \"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6\") " Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.416497 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" (UID: "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.443110 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-scripts" (OuterVolumeSpecName: "scripts") pod "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" (UID: "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.443219 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" (UID: "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.452582 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-kube-api-access-wpwvw" (OuterVolumeSpecName: "kube-api-access-wpwvw") pod "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" (UID: "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6"). InnerVolumeSpecName "kube-api-access-wpwvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.453066 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" (UID: "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.528335 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.528413 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpwvw\" (UniqueName: \"kubernetes.io/projected/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-kube-api-access-wpwvw\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.528428 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.528439 4992 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.528456 4992 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.528521 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5dfd9d5cfd-4mzm5"] Oct 04 04:54:07 crc kubenswrapper[4992]: E1004 04:54:07.529494 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" containerName="cinder-db-sync" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.529515 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" containerName="cinder-db-sync" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.529979 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" containerName="cinder-db-sync" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.533385 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.547255 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.547724 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.597686 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5dfd9d5cfd-4mzm5"] Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.615072 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-config-data" (OuterVolumeSpecName: "config-data") pod "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" (UID: "0a96c84d-f51b-48c0-9fa8-6040ed4f40f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.628225 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zcz9v" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.628743 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zcz9v" event={"ID":"0a96c84d-f51b-48c0-9fa8-6040ed4f40f6","Type":"ContainerDied","Data":"5b5568e537b1f23ebf4333c8dd7fa7944f5f2211b6e3965938bfe856d99094c0"} Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.628836 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b5568e537b1f23ebf4333c8dd7fa7944f5f2211b6e3965938bfe856d99094c0" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.630842 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-config-data\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.630928 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acb287d0-f746-4d74-acaa-4153228f3194-logs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.630965 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-public-tls-certs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.631045 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-config-data-custom\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.631145 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-internal-tls-certs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.631235 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnn4l\" (UniqueName: \"kubernetes.io/projected/acb287d0-f746-4d74-acaa-4153228f3194-kube-api-access-fnn4l\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.631298 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-combined-ca-bundle\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.631589 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.638892 4992 generic.go:334] "Generic (PLEG): container finished" podID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerID="301352d0463114eb0140352e1205933037b5790f271d2a12b10a4aeecaab85a7" exitCode=0 Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.640413 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652","Type":"ContainerDied","Data":"301352d0463114eb0140352e1205933037b5790f271d2a12b10a4aeecaab85a7"} Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.735273 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-config-data-custom\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.735846 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-internal-tls-certs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.735916 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnn4l\" (UniqueName: \"kubernetes.io/projected/acb287d0-f746-4d74-acaa-4153228f3194-kube-api-access-fnn4l\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.735961 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-combined-ca-bundle\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.736049 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-config-data\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.736123 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acb287d0-f746-4d74-acaa-4153228f3194-logs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.736155 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-public-tls-certs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.740570 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acb287d0-f746-4d74-acaa-4153228f3194-logs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.750234 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-config-data-custom\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.754793 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-public-tls-certs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.755014 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-config-data\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.760558 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-internal-tls-certs\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.768093 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acb287d0-f746-4d74-acaa-4153228f3194-combined-ca-bundle\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.772947 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnn4l\" (UniqueName: \"kubernetes.io/projected/acb287d0-f746-4d74-acaa-4153228f3194-kube-api-access-fnn4l\") pod \"barbican-api-5dfd9d5cfd-4mzm5\" (UID: \"acb287d0-f746-4d74-acaa-4153228f3194\") " pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.942831 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.944994 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.948963 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9h6dr" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.949344 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.949541 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.949821 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 04:54:07 crc kubenswrapper[4992]: I1004 04:54:07.964650 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.015862 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.042351 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.042495 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-scripts\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.042561 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.042602 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxtjc\" (UniqueName: \"kubernetes.io/projected/74a38db9-4407-4e43-8387-8d5214407154-kube-api-access-pxtjc\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.042656 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.042756 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74a38db9-4407-4e43-8387-8d5214407154-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.117406 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-dhd5n"] Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.143931 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-4pfdf"] Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.145874 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.148973 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.149072 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-scripts\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.149099 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.149132 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxtjc\" (UniqueName: \"kubernetes.io/projected/74a38db9-4407-4e43-8387-8d5214407154-kube-api-access-pxtjc\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.149164 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.149239 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74a38db9-4407-4e43-8387-8d5214407154-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.149424 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74a38db9-4407-4e43-8387-8d5214407154-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.163269 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.163533 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.164025 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-scripts\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.166692 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-4pfdf"] Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.170034 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.199139 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxtjc\" (UniqueName: \"kubernetes.io/projected/74a38db9-4407-4e43-8387-8d5214407154-kube-api-access-pxtjc\") pod \"cinder-scheduler-0\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.252464 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.252548 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.252649 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.252715 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtbfq\" (UniqueName: \"kubernetes.io/projected/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-kube-api-access-mtbfq\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.252771 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-config\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.252818 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.263909 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.264067 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.266555 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.270341 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.282582 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.308414 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.354901 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba54a486-1b98-4d22-af05-e5e723813c38-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.354953 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-scripts\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.354989 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba54a486-1b98-4d22-af05-e5e723813c38-logs\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355012 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hrhz\" (UniqueName: \"kubernetes.io/projected/ba54a486-1b98-4d22-af05-e5e723813c38-kube-api-access-6hrhz\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355046 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355082 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355131 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtbfq\" (UniqueName: \"kubernetes.io/projected/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-kube-api-access-mtbfq\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355155 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355173 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data-custom\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355226 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-config\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355288 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355306 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.355339 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.356313 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.356465 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.356803 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-config\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.357040 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.357188 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.375296 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtbfq\" (UniqueName: \"kubernetes.io/projected/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-kube-api-access-mtbfq\") pod \"dnsmasq-dns-69c986f6d7-4pfdf\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.458297 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba54a486-1b98-4d22-af05-e5e723813c38-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.458347 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-scripts\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.458430 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba54a486-1b98-4d22-af05-e5e723813c38-logs\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.458454 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hrhz\" (UniqueName: \"kubernetes.io/projected/ba54a486-1b98-4d22-af05-e5e723813c38-kube-api-access-6hrhz\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.458511 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.458525 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba54a486-1b98-4d22-af05-e5e723813c38-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.458554 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.458572 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data-custom\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.460297 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba54a486-1b98-4d22-af05-e5e723813c38-logs\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.463021 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data-custom\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.463992 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.464229 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-scripts\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.465607 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.477886 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hrhz\" (UniqueName: \"kubernetes.io/projected/ba54a486-1b98-4d22-af05-e5e723813c38-kube-api-access-6hrhz\") pod \"cinder-api-0\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.487483 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.587299 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:54:08 crc kubenswrapper[4992]: I1004 04:54:08.656025 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" podUID="8e07e645-fbe6-461a-96a2-027d28c6c993" containerName="dnsmasq-dns" containerID="cri-o://e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87" gracePeriod=10 Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.144504 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.276207 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fx6tl\" (UniqueName: \"kubernetes.io/projected/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-kube-api-access-fx6tl\") pod \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.276270 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-log-httpd\") pod \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.276326 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-config-data\") pod \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.276476 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-sg-core-conf-yaml\") pod \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.276557 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-run-httpd\") pod \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.276765 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-combined-ca-bundle\") pod \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.276805 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-scripts\") pod \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\" (UID: \"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.279029 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" (UID: "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.279793 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" (UID: "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.304152 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-scripts" (OuterVolumeSpecName: "scripts") pod "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" (UID: "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.312718 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-kube-api-access-fx6tl" (OuterVolumeSpecName: "kube-api-access-fx6tl") pod "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" (UID: "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652"). InnerVolumeSpecName "kube-api-access-fx6tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.359645 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.367099 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.385856 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.385903 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fx6tl\" (UniqueName: \"kubernetes.io/projected/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-kube-api-access-fx6tl\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.385917 4992 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.385929 4992 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: W1004 04:54:09.386347 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba54a486_1b98_4d22_af05_e5e723813c38.slice/crio-72bc890c28dab5985e445e3f6be1a5ff3f276929c573ac350297cf32565f096b WatchSource:0}: Error finding container 72bc890c28dab5985e445e3f6be1a5ff3f276929c573ac350297cf32565f096b: Status 404 returned error can't find the container with id 72bc890c28dab5985e445e3f6be1a5ff3f276929c573ac350297cf32565f096b Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.425517 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" (UID: "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.432541 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" (UID: "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.490089 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-config\") pod \"8e07e645-fbe6-461a-96a2-027d28c6c993\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.490268 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxzml\" (UniqueName: \"kubernetes.io/projected/8e07e645-fbe6-461a-96a2-027d28c6c993-kube-api-access-lxzml\") pod \"8e07e645-fbe6-461a-96a2-027d28c6c993\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.490316 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-nb\") pod \"8e07e645-fbe6-461a-96a2-027d28c6c993\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.490338 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-svc\") pod \"8e07e645-fbe6-461a-96a2-027d28c6c993\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.490461 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-sb\") pod \"8e07e645-fbe6-461a-96a2-027d28c6c993\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.490510 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-swift-storage-0\") pod \"8e07e645-fbe6-461a-96a2-027d28c6c993\" (UID: \"8e07e645-fbe6-461a-96a2-027d28c6c993\") " Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.491056 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.491072 4992 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.493739 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e07e645-fbe6-461a-96a2-027d28c6c993-kube-api-access-lxzml" (OuterVolumeSpecName: "kube-api-access-lxzml") pod "8e07e645-fbe6-461a-96a2-027d28c6c993" (UID: "8e07e645-fbe6-461a-96a2-027d28c6c993"). InnerVolumeSpecName "kube-api-access-lxzml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.523788 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-config-data" (OuterVolumeSpecName: "config-data") pod "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" (UID: "3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.562468 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.596500 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8e07e645-fbe6-461a-96a2-027d28c6c993" (UID: "8e07e645-fbe6-461a-96a2-027d28c6c993"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.598040 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxzml\" (UniqueName: \"kubernetes.io/projected/8e07e645-fbe6-461a-96a2-027d28c6c993-kube-api-access-lxzml\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.598109 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.598120 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.644776 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e07e645-fbe6-461a-96a2-027d28c6c993" (UID: "8e07e645-fbe6-461a-96a2-027d28c6c993"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.644850 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e07e645-fbe6-461a-96a2-027d28c6c993" (UID: "8e07e645-fbe6-461a-96a2-027d28c6c993"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.698345 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-config" (OuterVolumeSpecName: "config") pod "8e07e645-fbe6-461a-96a2-027d28c6c993" (UID: "8e07e645-fbe6-461a-96a2-027d28c6c993"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.698465 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e07e645-fbe6-461a-96a2-027d28c6c993" (UID: "8e07e645-fbe6-461a-96a2-027d28c6c993"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.719983 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.720032 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.720045 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.720053 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e07e645-fbe6-461a-96a2-027d28c6c993-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.721290 4992 generic.go:334] "Generic (PLEG): container finished" podID="8e07e645-fbe6-461a-96a2-027d28c6c993" containerID="e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87" exitCode=0 Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.721423 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" event={"ID":"8e07e645-fbe6-461a-96a2-027d28c6c993","Type":"ContainerDied","Data":"e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87"} Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.721459 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" event={"ID":"8e07e645-fbe6-461a-96a2-027d28c6c993","Type":"ContainerDied","Data":"ff53c5d5d7247afc72cc11977dcad12a18ca9fc3db12cffe9343e93135700b75"} Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.721481 4992 scope.go:117] "RemoveContainer" containerID="e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.723158 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-dhd5n" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.723206 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-4pfdf"] Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.724331 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ba54a486-1b98-4d22-af05-e5e723813c38","Type":"ContainerStarted","Data":"72bc890c28dab5985e445e3f6be1a5ff3f276929c573ac350297cf32565f096b"} Oct 04 04:54:09 crc kubenswrapper[4992]: W1004 04:54:09.738449 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b685d4a_92cb_4e60_acc1_dde941eaf3f3.slice/crio-53754fd6a4019d0d8e00c1171a98044b1e82522b716e5cf5a61560a566e621d6 WatchSource:0}: Error finding container 53754fd6a4019d0d8e00c1171a98044b1e82522b716e5cf5a61560a566e621d6: Status 404 returned error can't find the container with id 53754fd6a4019d0d8e00c1171a98044b1e82522b716e5cf5a61560a566e621d6 Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.739548 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652","Type":"ContainerDied","Data":"a10db9c31065f3b54aa484e7e6de2bd45aa81e67f7b56852c939936fd3336fcc"} Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.739722 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.742703 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" event={"ID":"a4b0448a-9b84-41c8-a87f-fbc17f92411b","Type":"ContainerStarted","Data":"77442411d03b936aaee8d7f95a6fe46e635b00192b784665cca8239975661c2c"} Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.742744 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" event={"ID":"a4b0448a-9b84-41c8-a87f-fbc17f92411b","Type":"ContainerStarted","Data":"0272b5585f3c00baa0cea0faa71cd1a00d3c58d34e7ecd9a01ca7db808ae65db"} Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.769830 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-775b66879-7sgzw" event={"ID":"70af60aa-3bee-44bf-874e-648aaefe032d","Type":"ContainerStarted","Data":"d8c9994f5ef31dbb8a22a4d7165e7f0b01f951081d9e69c35816ad4cae478c02"} Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.772147 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74a38db9-4407-4e43-8387-8d5214407154","Type":"ContainerStarted","Data":"da1e9fb193771d1b415e9575862b8f887ff570df6d6b629c03d755973416e233"} Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.777053 4992 scope.go:117] "RemoveContainer" containerID="70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.797216 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5dfd9d5cfd-4mzm5"] Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.818194 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5ff4b555c4-vn5np" podStartSLOduration=2.510303523 podStartE2EDuration="6.818173433s" podCreationTimestamp="2025-10-04 04:54:03 +0000 UTC" firstStartedPulling="2025-10-04 04:54:04.441390722 +0000 UTC m=+1278.289066190" lastFinishedPulling="2025-10-04 04:54:08.749260622 +0000 UTC m=+1282.596936100" observedRunningTime="2025-10-04 04:54:09.785410982 +0000 UTC m=+1283.633086460" watchObservedRunningTime="2025-10-04 04:54:09.818173433 +0000 UTC m=+1283.665848911" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.880827 4992 scope.go:117] "RemoveContainer" containerID="e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87" Oct 04 04:54:09 crc kubenswrapper[4992]: E1004 04:54:09.895769 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87\": container with ID starting with e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87 not found: ID does not exist" containerID="e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.895818 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87"} err="failed to get container status \"e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87\": rpc error: code = NotFound desc = could not find container \"e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87\": container with ID starting with e331555d9ee6963fd5d1998c670741b33ce55d19985260a217ecc00bc9520f87 not found: ID does not exist" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.895843 4992 scope.go:117] "RemoveContainer" containerID="70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746" Oct 04 04:54:09 crc kubenswrapper[4992]: E1004 04:54:09.896892 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746\": container with ID starting with 70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746 not found: ID does not exist" containerID="70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.896934 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746"} err="failed to get container status \"70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746\": rpc error: code = NotFound desc = could not find container \"70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746\": container with ID starting with 70a3223c87d2707e5b43fa74d861985d87f9e1441342db845fff5d1409971746 not found: ID does not exist" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.896960 4992 scope.go:117] "RemoveContainer" containerID="9fa907e890824e61aa2637b55000791843dfa8d47b50f4589871014f0b17f79d" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.921654 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.940837 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.959237 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-dhd5n"] Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.967547 4992 scope.go:117] "RemoveContainer" containerID="d6a75282c1df8dc25bc27d0c7fd5eeea686447968833ec972d161846603c177b" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.974734 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-dhd5n"] Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.991162 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:09 crc kubenswrapper[4992]: E1004 04:54:09.991842 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="sg-core" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.991866 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="sg-core" Oct 04 04:54:09 crc kubenswrapper[4992]: E1004 04:54:09.991880 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="ceilometer-notification-agent" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.991890 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="ceilometer-notification-agent" Oct 04 04:54:09 crc kubenswrapper[4992]: E1004 04:54:09.991910 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="proxy-httpd" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.991919 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="proxy-httpd" Oct 04 04:54:09 crc kubenswrapper[4992]: E1004 04:54:09.991964 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e07e645-fbe6-461a-96a2-027d28c6c993" containerName="dnsmasq-dns" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.991973 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e07e645-fbe6-461a-96a2-027d28c6c993" containerName="dnsmasq-dns" Oct 04 04:54:09 crc kubenswrapper[4992]: E1004 04:54:09.991990 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e07e645-fbe6-461a-96a2-027d28c6c993" containerName="init" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.991997 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e07e645-fbe6-461a-96a2-027d28c6c993" containerName="init" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.992696 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e07e645-fbe6-461a-96a2-027d28c6c993" containerName="dnsmasq-dns" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.992716 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="proxy-httpd" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.992736 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="sg-core" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.992784 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" containerName="ceilometer-notification-agent" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.996224 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.998567 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:54:09 crc kubenswrapper[4992]: I1004 04:54:09.998572 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.011844 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.023398 4992 scope.go:117] "RemoveContainer" containerID="301352d0463114eb0140352e1205933037b5790f271d2a12b10a4aeecaab85a7" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.129517 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-scripts\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.129649 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-run-httpd\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.129677 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-config-data\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.129697 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.129744 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rdk7\" (UniqueName: \"kubernetes.io/projected/933782b7-db86-44ca-97e9-34ae542598a4-kube-api-access-2rdk7\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.129866 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.129908 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-log-httpd\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.232799 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-log-httpd\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.233100 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-scripts\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.233179 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-run-httpd\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.233218 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-config-data\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.233243 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.233264 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rdk7\" (UniqueName: \"kubernetes.io/projected/933782b7-db86-44ca-97e9-34ae542598a4-kube-api-access-2rdk7\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.233276 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-log-httpd\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.233395 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.238418 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-run-httpd\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.238907 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-scripts\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.244311 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-config-data\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.245074 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.254810 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.255188 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rdk7\" (UniqueName: \"kubernetes.io/projected/933782b7-db86-44ca-97e9-34ae542598a4-kube-api-access-2rdk7\") pod \"ceilometer-0\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.329307 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.356613 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652" path="/var/lib/kubelet/pods/3e8b3ef0-9f96-4b4b-bb9c-bcab583a8652/volumes" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.357655 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e07e645-fbe6-461a-96a2-027d28c6c993" path="/var/lib/kubelet/pods/8e07e645-fbe6-461a-96a2-027d28c6c993/volumes" Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.854684 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-775b66879-7sgzw" event={"ID":"70af60aa-3bee-44bf-874e-648aaefe032d","Type":"ContainerStarted","Data":"b46dde0e2bf77afb377488e4f404ed7229dbae092d7ad63dea78e24514da0590"} Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.864130 4992 generic.go:334] "Generic (PLEG): container finished" podID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" containerID="90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482" exitCode=0 Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.864254 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" event={"ID":"2b685d4a-92cb-4e60-acc1-dde941eaf3f3","Type":"ContainerDied","Data":"90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482"} Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.864381 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" event={"ID":"2b685d4a-92cb-4e60-acc1-dde941eaf3f3","Type":"ContainerStarted","Data":"53754fd6a4019d0d8e00c1171a98044b1e82522b716e5cf5a61560a566e621d6"} Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.877141 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" event={"ID":"acb287d0-f746-4d74-acaa-4153228f3194","Type":"ContainerStarted","Data":"4038cfbe8cd6284a00434fa48a0c636858cdc64b716477547b19b30991496914"} Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.877193 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" event={"ID":"acb287d0-f746-4d74-acaa-4153228f3194","Type":"ContainerStarted","Data":"6b318d76740cc98906a0db39fdbe314406cdbb80e0cd620a9dc67ff12185a201"} Oct 04 04:54:10 crc kubenswrapper[4992]: I1004 04:54:10.886199 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-775b66879-7sgzw" podStartSLOduration=4.362227757 podStartE2EDuration="7.88615818s" podCreationTimestamp="2025-10-04 04:54:03 +0000 UTC" firstStartedPulling="2025-10-04 04:54:05.22574742 +0000 UTC m=+1279.073422888" lastFinishedPulling="2025-10-04 04:54:08.749677843 +0000 UTC m=+1282.597353311" observedRunningTime="2025-10-04 04:54:10.875818805 +0000 UTC m=+1284.723494293" watchObservedRunningTime="2025-10-04 04:54:10.88615818 +0000 UTC m=+1284.733833648" Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.386017 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.456206 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-55b5954cd6-68rnc" Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.517802 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.546152 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c6dc74865-2rgnq"] Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.546445 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c6dc74865-2rgnq" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon-log" containerID="cri-o://5360ba4db45400a267ecaceaa1c0924dabeb2f6d586b197d48d547360639404f" gracePeriod=30 Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.547067 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c6dc74865-2rgnq" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon" containerID="cri-o://0e838cf84f8b2286de7fee1771b85b59290fdf40b14d9269125d5379345a8111" gracePeriod=30 Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.942520 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74a38db9-4407-4e43-8387-8d5214407154","Type":"ContainerStarted","Data":"4ceeeff4f1f92a5821514a94ce5821dda1046ae29dc24b280e29082152cd9e78"} Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.961637 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerStarted","Data":"18e78b34097a5cdd3f444f2ee1c3d88525af6b6b173a3f87202346f36b082d02"} Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.964790 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ba54a486-1b98-4d22-af05-e5e723813c38","Type":"ContainerStarted","Data":"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde"} Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.967118 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" event={"ID":"2b685d4a-92cb-4e60-acc1-dde941eaf3f3","Type":"ContainerStarted","Data":"2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c"} Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.969102 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.981879 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" event={"ID":"acb287d0-f746-4d74-acaa-4153228f3194","Type":"ContainerStarted","Data":"06720d41025befee7291ff5b0011551f8da186c5381b55173362f668b6a44a7d"} Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.981933 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.981990 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:11 crc kubenswrapper[4992]: I1004 04:54:11.996455 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" podStartSLOduration=3.996429339 podStartE2EDuration="3.996429339s" podCreationTimestamp="2025-10-04 04:54:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:11.988508159 +0000 UTC m=+1285.836183637" watchObservedRunningTime="2025-10-04 04:54:11.996429339 +0000 UTC m=+1285.844104817" Oct 04 04:54:12 crc kubenswrapper[4992]: I1004 04:54:12.022444 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" podStartSLOduration=5.02242423 podStartE2EDuration="5.02242423s" podCreationTimestamp="2025-10-04 04:54:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:12.008480829 +0000 UTC m=+1285.856156307" watchObservedRunningTime="2025-10-04 04:54:12.02242423 +0000 UTC m=+1285.870099688" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.000482 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ba54a486-1b98-4d22-af05-e5e723813c38","Type":"ContainerStarted","Data":"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb"} Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.001089 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.000686 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" containerName="cinder-api" containerID="cri-o://3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb" gracePeriod=30 Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.002610 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" containerName="cinder-api-log" containerID="cri-o://36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde" gracePeriod=30 Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.004079 4992 generic.go:334] "Generic (PLEG): container finished" podID="eb31ac75-c92d-4b64-87b0-b5dfd194617b" containerID="d463686a8b7b1c4500d12116d647ebe0456ddb474581d901a3f5e2ea0ea04ccd" exitCode=0 Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.004173 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9pjhh" event={"ID":"eb31ac75-c92d-4b64-87b0-b5dfd194617b","Type":"ContainerDied","Data":"d463686a8b7b1c4500d12116d647ebe0456ddb474581d901a3f5e2ea0ea04ccd"} Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.009629 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74a38db9-4407-4e43-8387-8d5214407154","Type":"ContainerStarted","Data":"731b239e9bbed4a03b85666fa8bcd8a541eaa1c787ccf61c458fe5b483a976f8"} Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.015000 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerStarted","Data":"9ae04bc1b53e75d2f170ef9ccb3a4ea7ed93f4874c42f94844a1a1fb2c2b303b"} Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.031973 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.031953233 podStartE2EDuration="5.031953233s" podCreationTimestamp="2025-10-04 04:54:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:13.0280904 +0000 UTC m=+1286.875765868" watchObservedRunningTime="2025-10-04 04:54:13.031953233 +0000 UTC m=+1286.879628701" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.060285 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.109982707 podStartE2EDuration="6.060261235s" podCreationTimestamp="2025-10-04 04:54:07 +0000 UTC" firstStartedPulling="2025-10-04 04:54:09.582605152 +0000 UTC m=+1283.430280620" lastFinishedPulling="2025-10-04 04:54:10.53288368 +0000 UTC m=+1284.380559148" observedRunningTime="2025-10-04 04:54:13.050083775 +0000 UTC m=+1286.897759243" watchObservedRunningTime="2025-10-04 04:54:13.060261235 +0000 UTC m=+1286.907936873" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.313107 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.666144 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.751998 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-combined-ca-bundle\") pod \"ba54a486-1b98-4d22-af05-e5e723813c38\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.752117 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hrhz\" (UniqueName: \"kubernetes.io/projected/ba54a486-1b98-4d22-af05-e5e723813c38-kube-api-access-6hrhz\") pod \"ba54a486-1b98-4d22-af05-e5e723813c38\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.752169 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data\") pod \"ba54a486-1b98-4d22-af05-e5e723813c38\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.752225 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data-custom\") pod \"ba54a486-1b98-4d22-af05-e5e723813c38\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.752282 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba54a486-1b98-4d22-af05-e5e723813c38-etc-machine-id\") pod \"ba54a486-1b98-4d22-af05-e5e723813c38\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.752350 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba54a486-1b98-4d22-af05-e5e723813c38-logs\") pod \"ba54a486-1b98-4d22-af05-e5e723813c38\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.752401 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-scripts\") pod \"ba54a486-1b98-4d22-af05-e5e723813c38\" (UID: \"ba54a486-1b98-4d22-af05-e5e723813c38\") " Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.753196 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba54a486-1b98-4d22-af05-e5e723813c38-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ba54a486-1b98-4d22-af05-e5e723813c38" (UID: "ba54a486-1b98-4d22-af05-e5e723813c38"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.753612 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba54a486-1b98-4d22-af05-e5e723813c38-logs" (OuterVolumeSpecName: "logs") pod "ba54a486-1b98-4d22-af05-e5e723813c38" (UID: "ba54a486-1b98-4d22-af05-e5e723813c38"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.763444 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-scripts" (OuterVolumeSpecName: "scripts") pod "ba54a486-1b98-4d22-af05-e5e723813c38" (UID: "ba54a486-1b98-4d22-af05-e5e723813c38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.774661 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba54a486-1b98-4d22-af05-e5e723813c38-kube-api-access-6hrhz" (OuterVolumeSpecName: "kube-api-access-6hrhz") pod "ba54a486-1b98-4d22-af05-e5e723813c38" (UID: "ba54a486-1b98-4d22-af05-e5e723813c38"). InnerVolumeSpecName "kube-api-access-6hrhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.783707 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ba54a486-1b98-4d22-af05-e5e723813c38" (UID: "ba54a486-1b98-4d22-af05-e5e723813c38"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.826552 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba54a486-1b98-4d22-af05-e5e723813c38" (UID: "ba54a486-1b98-4d22-af05-e5e723813c38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.854042 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data" (OuterVolumeSpecName: "config-data") pod "ba54a486-1b98-4d22-af05-e5e723813c38" (UID: "ba54a486-1b98-4d22-af05-e5e723813c38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.854216 4992 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.854240 4992 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ba54a486-1b98-4d22-af05-e5e723813c38-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.854252 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba54a486-1b98-4d22-af05-e5e723813c38-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.854267 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.854279 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.854292 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hrhz\" (UniqueName: \"kubernetes.io/projected/ba54a486-1b98-4d22-af05-e5e723813c38-kube-api-access-6hrhz\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:13 crc kubenswrapper[4992]: I1004 04:54:13.854306 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba54a486-1b98-4d22-af05-e5e723813c38-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.031190 4992 generic.go:334] "Generic (PLEG): container finished" podID="ba54a486-1b98-4d22-af05-e5e723813c38" containerID="3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb" exitCode=0 Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.031530 4992 generic.go:334] "Generic (PLEG): container finished" podID="ba54a486-1b98-4d22-af05-e5e723813c38" containerID="36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde" exitCode=143 Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.031582 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ba54a486-1b98-4d22-af05-e5e723813c38","Type":"ContainerDied","Data":"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb"} Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.031614 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ba54a486-1b98-4d22-af05-e5e723813c38","Type":"ContainerDied","Data":"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde"} Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.031628 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ba54a486-1b98-4d22-af05-e5e723813c38","Type":"ContainerDied","Data":"72bc890c28dab5985e445e3f6be1a5ff3f276929c573ac350297cf32565f096b"} Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.031648 4992 scope.go:117] "RemoveContainer" containerID="3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.031786 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.037202 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerStarted","Data":"3e8fdcdab7c76f4ad8759bf84326581df5b6b4b5687ae287afffa25cae3e52eb"} Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.037249 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerStarted","Data":"44965bdbac199eba42f3c8cc171aeec494582286882e43a069cecbe2b513faa8"} Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.059442 4992 scope.go:117] "RemoveContainer" containerID="36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.088722 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.113616 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.117269 4992 scope.go:117] "RemoveContainer" containerID="3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb" Oct 04 04:54:14 crc kubenswrapper[4992]: E1004 04:54:14.118655 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb\": container with ID starting with 3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb not found: ID does not exist" containerID="3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.118720 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb"} err="failed to get container status \"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb\": rpc error: code = NotFound desc = could not find container \"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb\": container with ID starting with 3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb not found: ID does not exist" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.118759 4992 scope.go:117] "RemoveContainer" containerID="36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde" Oct 04 04:54:14 crc kubenswrapper[4992]: E1004 04:54:14.119644 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde\": container with ID starting with 36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde not found: ID does not exist" containerID="36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.119697 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde"} err="failed to get container status \"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde\": rpc error: code = NotFound desc = could not find container \"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde\": container with ID starting with 36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde not found: ID does not exist" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.119717 4992 scope.go:117] "RemoveContainer" containerID="3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.123331 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb"} err="failed to get container status \"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb\": rpc error: code = NotFound desc = could not find container \"3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb\": container with ID starting with 3c6703f32625634da7ac35a951df84919da4338727f593021bd7675351eff5cb not found: ID does not exist" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.123401 4992 scope.go:117] "RemoveContainer" containerID="36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.124066 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde"} err="failed to get container status \"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde\": rpc error: code = NotFound desc = could not find container \"36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde\": container with ID starting with 36bc13cb2b1678491d4796ec66f506e96994bc22291d8a97373a0186aa106dde not found: ID does not exist" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.145024 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:14 crc kubenswrapper[4992]: E1004 04:54:14.145417 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" containerName="cinder-api" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.145434 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" containerName="cinder-api" Oct 04 04:54:14 crc kubenswrapper[4992]: E1004 04:54:14.145460 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" containerName="cinder-api-log" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.145467 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" containerName="cinder-api-log" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.145621 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" containerName="cinder-api-log" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.145646 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" containerName="cinder-api" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.146565 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.151211 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.151325 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.152066 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.171977 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.262746 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-config-data-custom\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.262846 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.262895 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-scripts\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.262925 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.262958 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89482934-b40f-4685-948a-a2145549ba9e-logs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.263035 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-config-data\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.263136 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/89482934-b40f-4685-948a-a2145549ba9e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.263205 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5j5j\" (UniqueName: \"kubernetes.io/projected/89482934-b40f-4685-948a-a2145549ba9e-kube-api-access-n5j5j\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.263236 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.335463 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba54a486-1b98-4d22-af05-e5e723813c38" path="/var/lib/kubelet/pods/ba54a486-1b98-4d22-af05-e5e723813c38/volumes" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.360062 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364769 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/89482934-b40f-4685-948a-a2145549ba9e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364822 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5j5j\" (UniqueName: \"kubernetes.io/projected/89482934-b40f-4685-948a-a2145549ba9e-kube-api-access-n5j5j\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364839 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364875 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-config-data-custom\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364905 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364932 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-scripts\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364947 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364964 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89482934-b40f-4685-948a-a2145549ba9e-logs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.364994 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-config-data\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.370829 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/89482934-b40f-4685-948a-a2145549ba9e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.371701 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/89482934-b40f-4685-948a-a2145549ba9e-logs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.372003 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-config-data-custom\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.374803 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-scripts\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.375388 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.375495 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-config-data\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.377857 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.378116 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/89482934-b40f-4685-948a-a2145549ba9e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.391182 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5j5j\" (UniqueName: \"kubernetes.io/projected/89482934-b40f-4685-948a-a2145549ba9e-kube-api-access-n5j5j\") pod \"cinder-api-0\" (UID: \"89482934-b40f-4685-948a-a2145549ba9e\") " pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.466243 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-combined-ca-bundle\") pod \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.466307 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7bf7\" (UniqueName: \"kubernetes.io/projected/eb31ac75-c92d-4b64-87b0-b5dfd194617b-kube-api-access-c7bf7\") pod \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.466417 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-config\") pod \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\" (UID: \"eb31ac75-c92d-4b64-87b0-b5dfd194617b\") " Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.472877 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb31ac75-c92d-4b64-87b0-b5dfd194617b-kube-api-access-c7bf7" (OuterVolumeSpecName: "kube-api-access-c7bf7") pod "eb31ac75-c92d-4b64-87b0-b5dfd194617b" (UID: "eb31ac75-c92d-4b64-87b0-b5dfd194617b"). InnerVolumeSpecName "kube-api-access-c7bf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.491567 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.547069 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb31ac75-c92d-4b64-87b0-b5dfd194617b" (UID: "eb31ac75-c92d-4b64-87b0-b5dfd194617b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.547116 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-config" (OuterVolumeSpecName: "config") pod "eb31ac75-c92d-4b64-87b0-b5dfd194617b" (UID: "eb31ac75-c92d-4b64-87b0-b5dfd194617b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.568871 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.568911 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7bf7\" (UniqueName: \"kubernetes.io/projected/eb31ac75-c92d-4b64-87b0-b5dfd194617b-kube-api-access-c7bf7\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:14 crc kubenswrapper[4992]: I1004 04:54:14.568922 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb31ac75-c92d-4b64-87b0-b5dfd194617b-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.011281 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 04:54:15 crc kubenswrapper[4992]: W1004 04:54:15.033459 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89482934_b40f_4685_948a_a2145549ba9e.slice/crio-2840dec865bb49f855cfac8dca4540e96cf167fbdcc2c266ea73e87f9ad36df4 WatchSource:0}: Error finding container 2840dec865bb49f855cfac8dca4540e96cf167fbdcc2c266ea73e87f9ad36df4: Status 404 returned error can't find the container with id 2840dec865bb49f855cfac8dca4540e96cf167fbdcc2c266ea73e87f9ad36df4 Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.051704 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"89482934-b40f-4685-948a-a2145549ba9e","Type":"ContainerStarted","Data":"2840dec865bb49f855cfac8dca4540e96cf167fbdcc2c266ea73e87f9ad36df4"} Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.060776 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-9pjhh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.060690 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-9pjhh" event={"ID":"eb31ac75-c92d-4b64-87b0-b5dfd194617b","Type":"ContainerDied","Data":"dc8af903ef9e979b6f97cbce8cbe03feb7430493900c97a0703b8a3c5a2bb08f"} Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.062437 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc8af903ef9e979b6f97cbce8cbe03feb7430493900c97a0703b8a3c5a2bb08f" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.369810 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-4pfdf"] Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.400948 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jzvwf"] Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.404799 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" podUID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" containerName="dnsmasq-dns" containerID="cri-o://2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c" gracePeriod=10 Oct 04 04:54:15 crc kubenswrapper[4992]: E1004 04:54:15.408015 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb31ac75-c92d-4b64-87b0-b5dfd194617b" containerName="neutron-db-sync" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.408051 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb31ac75-c92d-4b64-87b0-b5dfd194617b" containerName="neutron-db-sync" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.408399 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb31ac75-c92d-4b64-87b0-b5dfd194617b" containerName="neutron-db-sync" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.409851 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.424693 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jzvwf"] Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.435030 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ddf7f7f7b-dtwqh"] Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.436965 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.439841 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.440153 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.440318 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.440511 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-czbnj" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.479501 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ddf7f7f7b-dtwqh"] Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.510776 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.510880 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-config\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.510922 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nl2l\" (UniqueName: \"kubernetes.io/projected/808727d5-8ae0-4be5-ad58-b01ab7bff16b-kube-api-access-4nl2l\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.510959 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.510990 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-svc\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.511012 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.615823 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-combined-ca-bundle\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.615930 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.616008 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-config\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.616539 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nl2l\" (UniqueName: \"kubernetes.io/projected/808727d5-8ae0-4be5-ad58-b01ab7bff16b-kube-api-access-4nl2l\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.616614 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-config\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.616668 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.616719 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-svc\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.616750 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.616843 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-httpd-config\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.616926 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhtj2\" (UniqueName: \"kubernetes.io/projected/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-kube-api-access-dhtj2\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.617024 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-ovndb-tls-certs\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.617407 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.621506 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-config\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.621696 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.622102 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.622754 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-svc\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.645089 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nl2l\" (UniqueName: \"kubernetes.io/projected/808727d5-8ae0-4be5-ad58-b01ab7bff16b-kube-api-access-4nl2l\") pod \"dnsmasq-dns-5784cf869f-jzvwf\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.719301 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhtj2\" (UniqueName: \"kubernetes.io/projected/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-kube-api-access-dhtj2\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.721642 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-ovndb-tls-certs\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.723003 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-combined-ca-bundle\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.724715 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-config\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.724933 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-httpd-config\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.745531 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-httpd-config\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.751746 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-config\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.752911 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhtj2\" (UniqueName: \"kubernetes.io/projected/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-kube-api-access-dhtj2\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.767145 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-combined-ca-bundle\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.767716 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-ovndb-tls-certs\") pod \"neutron-ddf7f7f7b-dtwqh\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.908210 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:15 crc kubenswrapper[4992]: I1004 04:54:15.917797 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.077898 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.079435 4992 generic.go:334] "Generic (PLEG): container finished" podID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" containerID="2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c" exitCode=0 Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.079574 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" event={"ID":"2b685d4a-92cb-4e60-acc1-dde941eaf3f3","Type":"ContainerDied","Data":"2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c"} Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.079665 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" event={"ID":"2b685d4a-92cb-4e60-acc1-dde941eaf3f3","Type":"ContainerDied","Data":"53754fd6a4019d0d8e00c1171a98044b1e82522b716e5cf5a61560a566e621d6"} Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.079741 4992 scope.go:117] "RemoveContainer" containerID="2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.120309 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerStarted","Data":"076434825bfa10ede425a2c4520f91bdb8f34d64169e9ae4b348d391bd9eb0db"} Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.122120 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.157260 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.447454117 podStartE2EDuration="7.15723504s" podCreationTimestamp="2025-10-04 04:54:09 +0000 UTC" firstStartedPulling="2025-10-04 04:54:11.408861492 +0000 UTC m=+1285.256536960" lastFinishedPulling="2025-10-04 04:54:15.118642415 +0000 UTC m=+1288.966317883" observedRunningTime="2025-10-04 04:54:16.151032455 +0000 UTC m=+1289.998707933" watchObservedRunningTime="2025-10-04 04:54:16.15723504 +0000 UTC m=+1290.004910508" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.166135 4992 scope.go:117] "RemoveContainer" containerID="90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.190762 4992 scope.go:117] "RemoveContainer" containerID="2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c" Oct 04 04:54:16 crc kubenswrapper[4992]: E1004 04:54:16.191290 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c\": container with ID starting with 2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c not found: ID does not exist" containerID="2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.191334 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c"} err="failed to get container status \"2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c\": rpc error: code = NotFound desc = could not find container \"2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c\": container with ID starting with 2d86076b1741f2078f90e072a2374625d3c047ff6fef464b570a617beb37733c not found: ID does not exist" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.191568 4992 scope.go:117] "RemoveContainer" containerID="90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482" Oct 04 04:54:16 crc kubenswrapper[4992]: E1004 04:54:16.191984 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482\": container with ID starting with 90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482 not found: ID does not exist" containerID="90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.192103 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482"} err="failed to get container status \"90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482\": rpc error: code = NotFound desc = could not find container \"90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482\": container with ID starting with 90e355fa81b257925fe51d02719bfc73c2b906af138e6d7a6243bc1d2d684482 not found: ID does not exist" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.248593 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-swift-storage-0\") pod \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.248660 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtbfq\" (UniqueName: \"kubernetes.io/projected/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-kube-api-access-mtbfq\") pod \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.248702 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-config\") pod \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.248783 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-sb\") pod \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.248851 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-nb\") pod \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.248873 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-svc\") pod \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\" (UID: \"2b685d4a-92cb-4e60-acc1-dde941eaf3f3\") " Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.281588 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-kube-api-access-mtbfq" (OuterVolumeSpecName: "kube-api-access-mtbfq") pod "2b685d4a-92cb-4e60-acc1-dde941eaf3f3" (UID: "2b685d4a-92cb-4e60-acc1-dde941eaf3f3"). InnerVolumeSpecName "kube-api-access-mtbfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.362020 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtbfq\" (UniqueName: \"kubernetes.io/projected/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-kube-api-access-mtbfq\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.422635 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-config" (OuterVolumeSpecName: "config") pod "2b685d4a-92cb-4e60-acc1-dde941eaf3f3" (UID: "2b685d4a-92cb-4e60-acc1-dde941eaf3f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.423379 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2b685d4a-92cb-4e60-acc1-dde941eaf3f3" (UID: "2b685d4a-92cb-4e60-acc1-dde941eaf3f3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.423649 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2b685d4a-92cb-4e60-acc1-dde941eaf3f3" (UID: "2b685d4a-92cb-4e60-acc1-dde941eaf3f3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.431511 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2b685d4a-92cb-4e60-acc1-dde941eaf3f3" (UID: "2b685d4a-92cb-4e60-acc1-dde941eaf3f3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.452668 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b685d4a-92cb-4e60-acc1-dde941eaf3f3" (UID: "2b685d4a-92cb-4e60-acc1-dde941eaf3f3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.463717 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.466632 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.466908 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.467055 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.467391 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b685d4a-92cb-4e60-acc1-dde941eaf3f3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.627167 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jzvwf"] Oct 04 04:54:16 crc kubenswrapper[4992]: I1004 04:54:16.723731 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ddf7f7f7b-dtwqh"] Oct 04 04:54:17 crc kubenswrapper[4992]: I1004 04:54:17.165699 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-4pfdf" Oct 04 04:54:17 crc kubenswrapper[4992]: I1004 04:54:17.170040 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"89482934-b40f-4685-948a-a2145549ba9e","Type":"ContainerStarted","Data":"c95969632bfcbc6726600359458b00fa43ba582493d52a07dca44b22d89ea66d"} Oct 04 04:54:17 crc kubenswrapper[4992]: I1004 04:54:17.180503 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ddf7f7f7b-dtwqh" event={"ID":"fffadcba-aa98-4d56-8b1f-86dc2435f8d0","Type":"ContainerStarted","Data":"2bd11cdc33b74ef68226026eac4a4594826f2ff40b6ab6ae2c497de6a236ecef"} Oct 04 04:54:17 crc kubenswrapper[4992]: I1004 04:54:17.192781 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" event={"ID":"808727d5-8ae0-4be5-ad58-b01ab7bff16b","Type":"ContainerStarted","Data":"1c74488a44757812d13b4a813739768401564eb189fc574a14be360198152dc8"} Oct 04 04:54:17 crc kubenswrapper[4992]: I1004 04:54:17.219352 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-4pfdf"] Oct 04 04:54:17 crc kubenswrapper[4992]: I1004 04:54:17.227987 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-4pfdf"] Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.024731 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.075885 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.201441 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ddf7f7f7b-dtwqh" event={"ID":"fffadcba-aa98-4d56-8b1f-86dc2435f8d0","Type":"ContainerStarted","Data":"29c2ed4603c34fd34e9b934e9245b45af9e552ee72c0bb1b0366afc2cbac49ee"} Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.201683 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ddf7f7f7b-dtwqh" event={"ID":"fffadcba-aa98-4d56-8b1f-86dc2435f8d0","Type":"ContainerStarted","Data":"fc353a7dda606f02c0bd0b3e4311fc9621bc9b49fc831576d3ec57048058ed9c"} Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.201729 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.203118 4992 generic.go:334] "Generic (PLEG): container finished" podID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerID="4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07" exitCode=0 Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.203179 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" event={"ID":"808727d5-8ae0-4be5-ad58-b01ab7bff16b","Type":"ContainerDied","Data":"4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07"} Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.216630 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"89482934-b40f-4685-948a-a2145549ba9e","Type":"ContainerStarted","Data":"3b895922393a8d5b9f4e37fb2561170e4e79fe28b49ddca1a68ab754a6cd646d"} Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.216678 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.225242 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-ddf7f7f7b-dtwqh" podStartSLOduration=3.225225397 podStartE2EDuration="3.225225397s" podCreationTimestamp="2025-10-04 04:54:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:18.224168228 +0000 UTC m=+1292.071843696" watchObservedRunningTime="2025-10-04 04:54:18.225225397 +0000 UTC m=+1292.072900865" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.258751 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.258730967 podStartE2EDuration="4.258730967s" podCreationTimestamp="2025-10-04 04:54:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:18.249841881 +0000 UTC m=+1292.097517359" watchObservedRunningTime="2025-10-04 04:54:18.258730967 +0000 UTC m=+1292.106406435" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.355314 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" path="/var/lib/kubelet/pods/2b685d4a-92cb-4e60-acc1-dde941eaf3f3/volumes" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.512927 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8699c8bbf-h2hwx"] Oct 04 04:54:18 crc kubenswrapper[4992]: E1004 04:54:18.513743 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" containerName="init" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.513766 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" containerName="init" Oct 04 04:54:18 crc kubenswrapper[4992]: E1004 04:54:18.513807 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" containerName="dnsmasq-dns" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.513814 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" containerName="dnsmasq-dns" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.514017 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b685d4a-92cb-4e60-acc1-dde941eaf3f3" containerName="dnsmasq-dns" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.514980 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.534416 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8699c8bbf-h2hwx"] Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.537042 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.537174 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.637544 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-ovndb-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.637848 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-config\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.638019 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-public-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.638179 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-httpd-config\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.638336 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-internal-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.638517 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwc8v\" (UniqueName: \"kubernetes.io/projected/404db73d-c716-4d1a-808b-92b193c9b7cf-kube-api-access-pwc8v\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.638638 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-combined-ca-bundle\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.689811 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.736638 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.740483 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwc8v\" (UniqueName: \"kubernetes.io/projected/404db73d-c716-4d1a-808b-92b193c9b7cf-kube-api-access-pwc8v\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.740762 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-combined-ca-bundle\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.740886 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-ovndb-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.741001 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-config\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.741152 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-public-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.741306 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-httpd-config\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.741463 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-internal-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.867179 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-internal-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.868340 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-config\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.875036 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-combined-ca-bundle\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.877165 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-public-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.888509 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-httpd-config\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.890671 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/404db73d-c716-4d1a-808b-92b193c9b7cf-ovndb-tls-certs\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:18 crc kubenswrapper[4992]: I1004 04:54:18.894981 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwc8v\" (UniqueName: \"kubernetes.io/projected/404db73d-c716-4d1a-808b-92b193c9b7cf-kube-api-access-pwc8v\") pod \"neutron-8699c8bbf-h2hwx\" (UID: \"404db73d-c716-4d1a-808b-92b193c9b7cf\") " pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:19 crc kubenswrapper[4992]: I1004 04:54:19.175917 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:19 crc kubenswrapper[4992]: I1004 04:54:19.238446 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" event={"ID":"808727d5-8ae0-4be5-ad58-b01ab7bff16b","Type":"ContainerStarted","Data":"c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84"} Oct 04 04:54:19 crc kubenswrapper[4992]: I1004 04:54:19.239095 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="74a38db9-4407-4e43-8387-8d5214407154" containerName="cinder-scheduler" containerID="cri-o://4ceeeff4f1f92a5821514a94ce5821dda1046ae29dc24b280e29082152cd9e78" gracePeriod=30 Oct 04 04:54:19 crc kubenswrapper[4992]: I1004 04:54:19.239208 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="74a38db9-4407-4e43-8387-8d5214407154" containerName="probe" containerID="cri-o://731b239e9bbed4a03b85666fa8bcd8a541eaa1c787ccf61c458fe5b483a976f8" gracePeriod=30 Oct 04 04:54:19 crc kubenswrapper[4992]: I1004 04:54:19.280788 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" podStartSLOduration=4.280765761 podStartE2EDuration="4.280765761s" podCreationTimestamp="2025-10-04 04:54:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:19.267906459 +0000 UTC m=+1293.115581927" watchObservedRunningTime="2025-10-04 04:54:19.280765761 +0000 UTC m=+1293.128441229" Oct 04 04:54:19 crc kubenswrapper[4992]: I1004 04:54:19.793250 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8699c8bbf-h2hwx"] Oct 04 04:54:19 crc kubenswrapper[4992]: I1004 04:54:19.958632 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.270060 4992 generic.go:334] "Generic (PLEG): container finished" podID="94c198d3-0808-4077-951a-fe14a94409ea" containerID="0e838cf84f8b2286de7fee1771b85b59290fdf40b14d9269125d5379345a8111" exitCode=1 Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.270444 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6dc74865-2rgnq" event={"ID":"94c198d3-0808-4077-951a-fe14a94409ea","Type":"ContainerDied","Data":"0e838cf84f8b2286de7fee1771b85b59290fdf40b14d9269125d5379345a8111"} Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.274131 4992 generic.go:334] "Generic (PLEG): container finished" podID="74a38db9-4407-4e43-8387-8d5214407154" containerID="731b239e9bbed4a03b85666fa8bcd8a541eaa1c787ccf61c458fe5b483a976f8" exitCode=0 Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.274191 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74a38db9-4407-4e43-8387-8d5214407154","Type":"ContainerDied","Data":"731b239e9bbed4a03b85666fa8bcd8a541eaa1c787ccf61c458fe5b483a976f8"} Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.276464 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8699c8bbf-h2hwx" event={"ID":"404db73d-c716-4d1a-808b-92b193c9b7cf","Type":"ContainerStarted","Data":"20c835300ca5eae05234d578b708a78d3e4989c5eb3edd3f625ab7f76cc9f033"} Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.276518 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8699c8bbf-h2hwx" event={"ID":"404db73d-c716-4d1a-808b-92b193c9b7cf","Type":"ContainerStarted","Data":"846dfa31e3bb36316fd8696a9bcdd3daf59c60251f23872402be281d0b99f35f"} Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.276576 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.720384 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5dfd9d5cfd-4mzm5" Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.792187 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-bcc7d5cbb-phnsf"] Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.792526 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-bcc7d5cbb-phnsf" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api-log" containerID="cri-o://702f1d81b231f96c38514f0e51d7f53cdd3cafbd3878d84a9a298ed152f1ef87" gracePeriod=30 Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.793010 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-bcc7d5cbb-phnsf" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api" containerID="cri-o://d3a0eb567bc650c5d526e462af54066bd733518d0f0d841e5eaf06d0521a1fed" gracePeriod=30 Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.899517 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.899645 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.899727 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.906212 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c66b07d0127863fb32b4550ae384badc1dd16f486eaff1e71f5b9fa6a2b2449"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:54:20 crc kubenswrapper[4992]: I1004 04:54:20.906300 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://5c66b07d0127863fb32b4550ae384badc1dd16f486eaff1e71f5b9fa6a2b2449" gracePeriod=600 Oct 04 04:54:21 crc kubenswrapper[4992]: I1004 04:54:21.287325 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8699c8bbf-h2hwx" event={"ID":"404db73d-c716-4d1a-808b-92b193c9b7cf","Type":"ContainerStarted","Data":"51f63353f62502a211952075ab0d43a17101c10fb71ce81c54d0fb4b65fbe900"} Oct 04 04:54:21 crc kubenswrapper[4992]: I1004 04:54:21.287705 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:21 crc kubenswrapper[4992]: I1004 04:54:21.291816 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="5c66b07d0127863fb32b4550ae384badc1dd16f486eaff1e71f5b9fa6a2b2449" exitCode=0 Oct 04 04:54:21 crc kubenswrapper[4992]: I1004 04:54:21.291900 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"5c66b07d0127863fb32b4550ae384badc1dd16f486eaff1e71f5b9fa6a2b2449"} Oct 04 04:54:21 crc kubenswrapper[4992]: I1004 04:54:21.291937 4992 scope.go:117] "RemoveContainer" containerID="bb88432e3860828bfc70223a5d4820bc8d7427ffafdb37bf5816fc3fffccf08d" Oct 04 04:54:21 crc kubenswrapper[4992]: I1004 04:54:21.299739 4992 generic.go:334] "Generic (PLEG): container finished" podID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerID="702f1d81b231f96c38514f0e51d7f53cdd3cafbd3878d84a9a298ed152f1ef87" exitCode=143 Oct 04 04:54:21 crc kubenswrapper[4992]: I1004 04:54:21.299783 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcc7d5cbb-phnsf" event={"ID":"57dc2121-ed97-4dea-903f-6d3b92b539ff","Type":"ContainerDied","Data":"702f1d81b231f96c38514f0e51d7f53cdd3cafbd3878d84a9a298ed152f1ef87"} Oct 04 04:54:21 crc kubenswrapper[4992]: I1004 04:54:21.321260 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8699c8bbf-h2hwx" podStartSLOduration=3.321240246 podStartE2EDuration="3.321240246s" podCreationTimestamp="2025-10-04 04:54:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:21.311927759 +0000 UTC m=+1295.159603227" watchObservedRunningTime="2025-10-04 04:54:21.321240246 +0000 UTC m=+1295.168915714" Oct 04 04:54:22 crc kubenswrapper[4992]: I1004 04:54:22.331808 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"b7858b04a08dbf3c2f36756b5a0b3054268a106fb6be0797a45494ff8fb3a8ad"} Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.340855 4992 generic.go:334] "Generic (PLEG): container finished" podID="74a38db9-4407-4e43-8387-8d5214407154" containerID="4ceeeff4f1f92a5821514a94ce5821dda1046ae29dc24b280e29082152cd9e78" exitCode=0 Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.340946 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74a38db9-4407-4e43-8387-8d5214407154","Type":"ContainerDied","Data":"4ceeeff4f1f92a5821514a94ce5821dda1046ae29dc24b280e29082152cd9e78"} Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.604109 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.647011 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-scripts\") pod \"74a38db9-4407-4e43-8387-8d5214407154\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.647451 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data-custom\") pod \"74a38db9-4407-4e43-8387-8d5214407154\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.647478 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74a38db9-4407-4e43-8387-8d5214407154-etc-machine-id\") pod \"74a38db9-4407-4e43-8387-8d5214407154\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.647695 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data\") pod \"74a38db9-4407-4e43-8387-8d5214407154\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.647742 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-combined-ca-bundle\") pod \"74a38db9-4407-4e43-8387-8d5214407154\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.647823 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxtjc\" (UniqueName: \"kubernetes.io/projected/74a38db9-4407-4e43-8387-8d5214407154-kube-api-access-pxtjc\") pod \"74a38db9-4407-4e43-8387-8d5214407154\" (UID: \"74a38db9-4407-4e43-8387-8d5214407154\") " Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.648476 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74a38db9-4407-4e43-8387-8d5214407154-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "74a38db9-4407-4e43-8387-8d5214407154" (UID: "74a38db9-4407-4e43-8387-8d5214407154"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.659656 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-scripts" (OuterVolumeSpecName: "scripts") pod "74a38db9-4407-4e43-8387-8d5214407154" (UID: "74a38db9-4407-4e43-8387-8d5214407154"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.660063 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "74a38db9-4407-4e43-8387-8d5214407154" (UID: "74a38db9-4407-4e43-8387-8d5214407154"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.678552 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a38db9-4407-4e43-8387-8d5214407154-kube-api-access-pxtjc" (OuterVolumeSpecName: "kube-api-access-pxtjc") pod "74a38db9-4407-4e43-8387-8d5214407154" (UID: "74a38db9-4407-4e43-8387-8d5214407154"). InnerVolumeSpecName "kube-api-access-pxtjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.750374 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxtjc\" (UniqueName: \"kubernetes.io/projected/74a38db9-4407-4e43-8387-8d5214407154-kube-api-access-pxtjc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.750414 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.750427 4992 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.750438 4992 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/74a38db9-4407-4e43-8387-8d5214407154-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.752066 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74a38db9-4407-4e43-8387-8d5214407154" (UID: "74a38db9-4407-4e43-8387-8d5214407154"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.815511 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data" (OuterVolumeSpecName: "config-data") pod "74a38db9-4407-4e43-8387-8d5214407154" (UID: "74a38db9-4407-4e43-8387-8d5214407154"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.851910 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:23 crc kubenswrapper[4992]: I1004 04:54:23.851937 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a38db9-4407-4e43-8387-8d5214407154-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.220927 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bcc7d5cbb-phnsf" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": dial tcp 10.217.0.156:9311: connect: connection refused" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.220933 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bcc7d5cbb-phnsf" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": dial tcp 10.217.0.156:9311: connect: connection refused" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.366086 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.370606 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"74a38db9-4407-4e43-8387-8d5214407154","Type":"ContainerDied","Data":"da1e9fb193771d1b415e9575862b8f887ff570df6d6b629c03d755973416e233"} Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.370669 4992 scope.go:117] "RemoveContainer" containerID="731b239e9bbed4a03b85666fa8bcd8a541eaa1c787ccf61c458fe5b483a976f8" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.384135 4992 generic.go:334] "Generic (PLEG): container finished" podID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerID="d3a0eb567bc650c5d526e462af54066bd733518d0f0d841e5eaf06d0521a1fed" exitCode=0 Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.384392 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcc7d5cbb-phnsf" event={"ID":"57dc2121-ed97-4dea-903f-6d3b92b539ff","Type":"ContainerDied","Data":"d3a0eb567bc650c5d526e462af54066bd733518d0f0d841e5eaf06d0521a1fed"} Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.405679 4992 scope.go:117] "RemoveContainer" containerID="4ceeeff4f1f92a5821514a94ce5821dda1046ae29dc24b280e29082152cd9e78" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.414643 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.425124 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.443904 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:24 crc kubenswrapper[4992]: E1004 04:54:24.444485 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a38db9-4407-4e43-8387-8d5214407154" containerName="cinder-scheduler" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.444558 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a38db9-4407-4e43-8387-8d5214407154" containerName="cinder-scheduler" Oct 04 04:54:24 crc kubenswrapper[4992]: E1004 04:54:24.444641 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a38db9-4407-4e43-8387-8d5214407154" containerName="probe" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.444691 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a38db9-4407-4e43-8387-8d5214407154" containerName="probe" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.444930 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a38db9-4407-4e43-8387-8d5214407154" containerName="probe" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.445275 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a38db9-4407-4e43-8387-8d5214407154" containerName="cinder-scheduler" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.458741 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.462732 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.474715 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.576738 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.576889 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.576925 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qptnm\" (UniqueName: \"kubernetes.io/projected/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-kube-api-access-qptnm\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.577002 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-scripts\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.577074 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-config-data\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.577182 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.645085 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.678511 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-combined-ca-bundle\") pod \"57dc2121-ed97-4dea-903f-6d3b92b539ff\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.678575 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dc2121-ed97-4dea-903f-6d3b92b539ff-logs\") pod \"57dc2121-ed97-4dea-903f-6d3b92b539ff\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.678608 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data\") pod \"57dc2121-ed97-4dea-903f-6d3b92b539ff\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.678755 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data-custom\") pod \"57dc2121-ed97-4dea-903f-6d3b92b539ff\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.678833 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79rzn\" (UniqueName: \"kubernetes.io/projected/57dc2121-ed97-4dea-903f-6d3b92b539ff-kube-api-access-79rzn\") pod \"57dc2121-ed97-4dea-903f-6d3b92b539ff\" (UID: \"57dc2121-ed97-4dea-903f-6d3b92b539ff\") " Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.679015 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.679043 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qptnm\" (UniqueName: \"kubernetes.io/projected/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-kube-api-access-qptnm\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.679094 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-scripts\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.679135 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-config-data\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.679176 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.679194 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.679401 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.680159 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57dc2121-ed97-4dea-903f-6d3b92b539ff-logs" (OuterVolumeSpecName: "logs") pod "57dc2121-ed97-4dea-903f-6d3b92b539ff" (UID: "57dc2121-ed97-4dea-903f-6d3b92b539ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.686533 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.687338 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "57dc2121-ed97-4dea-903f-6d3b92b539ff" (UID: "57dc2121-ed97-4dea-903f-6d3b92b539ff"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.688585 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57dc2121-ed97-4dea-903f-6d3b92b539ff-kube-api-access-79rzn" (OuterVolumeSpecName: "kube-api-access-79rzn") pod "57dc2121-ed97-4dea-903f-6d3b92b539ff" (UID: "57dc2121-ed97-4dea-903f-6d3b92b539ff"). InnerVolumeSpecName "kube-api-access-79rzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.702690 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.705398 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-scripts\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.705417 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-config-data\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.712089 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qptnm\" (UniqueName: \"kubernetes.io/projected/6d24bc2b-9416-4bf0-aff1-081826b9c9d9-kube-api-access-qptnm\") pod \"cinder-scheduler-0\" (UID: \"6d24bc2b-9416-4bf0-aff1-081826b9c9d9\") " pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.747516 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57dc2121-ed97-4dea-903f-6d3b92b539ff" (UID: "57dc2121-ed97-4dea-903f-6d3b92b539ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.773513 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data" (OuterVolumeSpecName: "config-data") pod "57dc2121-ed97-4dea-903f-6d3b92b539ff" (UID: "57dc2121-ed97-4dea-903f-6d3b92b539ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.782086 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79rzn\" (UniqueName: \"kubernetes.io/projected/57dc2121-ed97-4dea-903f-6d3b92b539ff-kube-api-access-79rzn\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.782115 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.782124 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57dc2121-ed97-4dea-903f-6d3b92b539ff-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.782138 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.782146 4992 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57dc2121-ed97-4dea-903f-6d3b92b539ff-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.800972 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 04:54:24 crc kubenswrapper[4992]: I1004 04:54:24.976816 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5bb778b4c-snbm9" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.327736 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 04:54:25 crc kubenswrapper[4992]: W1004 04:54:25.330878 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d24bc2b_9416_4bf0_aff1_081826b9c9d9.slice/crio-415d63b0499cae5a8b7b90bdcdc4318c09672787500fa0937f6fb00c6f51950b WatchSource:0}: Error finding container 415d63b0499cae5a8b7b90bdcdc4318c09672787500fa0937f6fb00c6f51950b: Status 404 returned error can't find the container with id 415d63b0499cae5a8b7b90bdcdc4318c09672787500fa0937f6fb00c6f51950b Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.397017 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6d24bc2b-9416-4bf0-aff1-081826b9c9d9","Type":"ContainerStarted","Data":"415d63b0499cae5a8b7b90bdcdc4318c09672787500fa0937f6fb00c6f51950b"} Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.400452 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcc7d5cbb-phnsf" event={"ID":"57dc2121-ed97-4dea-903f-6d3b92b539ff","Type":"ContainerDied","Data":"3aab48c089737e05c712dfde2e52b1d3f10d72ff94a50dc4108fd05c4a25156a"} Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.400484 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bcc7d5cbb-phnsf" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.400493 4992 scope.go:117] "RemoveContainer" containerID="d3a0eb567bc650c5d526e462af54066bd733518d0f0d841e5eaf06d0521a1fed" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.541878 4992 scope.go:117] "RemoveContainer" containerID="702f1d81b231f96c38514f0e51d7f53cdd3cafbd3878d84a9a298ed152f1ef87" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.551691 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-bcc7d5cbb-phnsf"] Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.572015 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-bcc7d5cbb-phnsf"] Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.802167 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 04:54:25 crc kubenswrapper[4992]: E1004 04:54:25.802570 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.802589 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api" Oct 04 04:54:25 crc kubenswrapper[4992]: E1004 04:54:25.802618 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api-log" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.802625 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api-log" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.807574 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api-log" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.807613 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" containerName="barbican-api" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.808260 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.816127 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-8prmt" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.816432 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.818886 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.829503 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.904798 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.904875 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.904900 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.904923 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gtbk\" (UniqueName: \"kubernetes.io/projected/f5fac4a9-5b82-4908-a798-69469297f7e6-kube-api-access-7gtbk\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.916676 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.979883 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-qs2sd"] Oct 04 04:54:25 crc kubenswrapper[4992]: I1004 04:54:25.980434 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" podUID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerName="dnsmasq-dns" containerID="cri-o://48780c5b8756483e0bceadf4f9d0c6c2d167eade4004c20641ca7c79af35cbb5" gracePeriod=10 Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.011527 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.011664 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.011703 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.011762 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gtbk\" (UniqueName: \"kubernetes.io/projected/f5fac4a9-5b82-4908-a798-69469297f7e6-kube-api-access-7gtbk\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.013577 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.018527 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.018956 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.042831 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gtbk\" (UniqueName: \"kubernetes.io/projected/f5fac4a9-5b82-4908-a798-69469297f7e6-kube-api-access-7gtbk\") pod \"openstackclient\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.076220 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.076874 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.087243 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.110438 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.114858 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.128961 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.214436 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvfjl\" (UniqueName: \"kubernetes.io/projected/8a864935-07eb-48c7-a02b-a81ff050ce4b-kube-api-access-cvfjl\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.214530 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a864935-07eb-48c7-a02b-a81ff050ce4b-openstack-config-secret\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.214599 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a864935-07eb-48c7-a02b-a81ff050ce4b-openstack-config\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.214713 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a864935-07eb-48c7-a02b-a81ff050ce4b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: E1004 04:54:26.290944 4992 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 04 04:54:26 crc kubenswrapper[4992]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_f5fac4a9-5b82-4908-a798-69469297f7e6_0(2c0fa31ae2a9ead54e3042d8f5bd0d01b20b796144cf015339cd16c479c745ce): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"2c0fa31ae2a9ead54e3042d8f5bd0d01b20b796144cf015339cd16c479c745ce" Netns:"/var/run/netns/c89ed507-c8d2-451f-a5fd-3a353ede4520" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=2c0fa31ae2a9ead54e3042d8f5bd0d01b20b796144cf015339cd16c479c745ce;K8S_POD_UID=f5fac4a9-5b82-4908-a798-69469297f7e6" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/f5fac4a9-5b82-4908-a798-69469297f7e6]: expected pod UID "f5fac4a9-5b82-4908-a798-69469297f7e6" but got "8a864935-07eb-48c7-a02b-a81ff050ce4b" from Kube API Oct 04 04:54:26 crc kubenswrapper[4992]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 04 04:54:26 crc kubenswrapper[4992]: > Oct 04 04:54:26 crc kubenswrapper[4992]: E1004 04:54:26.291342 4992 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 04 04:54:26 crc kubenswrapper[4992]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_f5fac4a9-5b82-4908-a798-69469297f7e6_0(2c0fa31ae2a9ead54e3042d8f5bd0d01b20b796144cf015339cd16c479c745ce): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"2c0fa31ae2a9ead54e3042d8f5bd0d01b20b796144cf015339cd16c479c745ce" Netns:"/var/run/netns/c89ed507-c8d2-451f-a5fd-3a353ede4520" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=2c0fa31ae2a9ead54e3042d8f5bd0d01b20b796144cf015339cd16c479c745ce;K8S_POD_UID=f5fac4a9-5b82-4908-a798-69469297f7e6" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/f5fac4a9-5b82-4908-a798-69469297f7e6]: expected pod UID "f5fac4a9-5b82-4908-a798-69469297f7e6" but got "8a864935-07eb-48c7-a02b-a81ff050ce4b" from Kube API Oct 04 04:54:26 crc kubenswrapper[4992]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 04 04:54:26 crc kubenswrapper[4992]: > pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.317526 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvfjl\" (UniqueName: \"kubernetes.io/projected/8a864935-07eb-48c7-a02b-a81ff050ce4b-kube-api-access-cvfjl\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.317581 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a864935-07eb-48c7-a02b-a81ff050ce4b-openstack-config-secret\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.317660 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a864935-07eb-48c7-a02b-a81ff050ce4b-openstack-config\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.317776 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a864935-07eb-48c7-a02b-a81ff050ce4b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.319143 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8a864935-07eb-48c7-a02b-a81ff050ce4b-openstack-config\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.321920 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" podUID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.144:5353: connect: connection refused" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.324987 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a864935-07eb-48c7-a02b-a81ff050ce4b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.327265 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8a864935-07eb-48c7-a02b-a81ff050ce4b-openstack-config-secret\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.338019 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvfjl\" (UniqueName: \"kubernetes.io/projected/8a864935-07eb-48c7-a02b-a81ff050ce4b-kube-api-access-cvfjl\") pod \"openstackclient\" (UID: \"8a864935-07eb-48c7-a02b-a81ff050ce4b\") " pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.354522 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57dc2121-ed97-4dea-903f-6d3b92b539ff" path="/var/lib/kubelet/pods/57dc2121-ed97-4dea-903f-6d3b92b539ff/volumes" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.355262 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a38db9-4407-4e43-8387-8d5214407154" path="/var/lib/kubelet/pods/74a38db9-4407-4e43-8387-8d5214407154/volumes" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.438723 4992 generic.go:334] "Generic (PLEG): container finished" podID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerID="48780c5b8756483e0bceadf4f9d0c6c2d167eade4004c20641ca7c79af35cbb5" exitCode=0 Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.438778 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" event={"ID":"8e8eb649-79cf-41a7-a2be-a0d713c6d741","Type":"ContainerDied","Data":"48780c5b8756483e0bceadf4f9d0c6c2d167eade4004c20641ca7c79af35cbb5"} Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.452341 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.452534 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6d24bc2b-9416-4bf0-aff1-081826b9c9d9","Type":"ContainerStarted","Data":"588cd67a06c1ba3fbc54d80c79d7447acffd47584775e385023f03a07bbcc77c"} Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.474074 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.478730 4992 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="f5fac4a9-5b82-4908-a798-69469297f7e6" podUID="8a864935-07eb-48c7-a02b-a81ff050ce4b" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.523182 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gtbk\" (UniqueName: \"kubernetes.io/projected/f5fac4a9-5b82-4908-a798-69469297f7e6-kube-api-access-7gtbk\") pod \"f5fac4a9-5b82-4908-a798-69469297f7e6\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.523498 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-combined-ca-bundle\") pod \"f5fac4a9-5b82-4908-a798-69469297f7e6\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.523642 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config-secret\") pod \"f5fac4a9-5b82-4908-a798-69469297f7e6\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.523708 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config\") pod \"f5fac4a9-5b82-4908-a798-69469297f7e6\" (UID: \"f5fac4a9-5b82-4908-a798-69469297f7e6\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.524495 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "f5fac4a9-5b82-4908-a798-69469297f7e6" (UID: "f5fac4a9-5b82-4908-a798-69469297f7e6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.535552 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5fac4a9-5b82-4908-a798-69469297f7e6-kube-api-access-7gtbk" (OuterVolumeSpecName: "kube-api-access-7gtbk") pod "f5fac4a9-5b82-4908-a798-69469297f7e6" (UID: "f5fac4a9-5b82-4908-a798-69469297f7e6"). InnerVolumeSpecName "kube-api-access-7gtbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.543154 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "f5fac4a9-5b82-4908-a798-69469297f7e6" (UID: "f5fac4a9-5b82-4908-a798-69469297f7e6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.546130 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5fac4a9-5b82-4908-a798-69469297f7e6" (UID: "f5fac4a9-5b82-4908-a798-69469297f7e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.553327 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.625877 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.625910 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5fac4a9-5b82-4908-a798-69469297f7e6-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.625934 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gtbk\" (UniqueName: \"kubernetes.io/projected/f5fac4a9-5b82-4908-a798-69469297f7e6-kube-api-access-7gtbk\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.625945 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5fac4a9-5b82-4908-a798-69469297f7e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.697663 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.829965 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-svc\") pod \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.830068 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-config\") pod \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.830103 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-nb\") pod \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.830227 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-sb\") pod \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.830255 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-swift-storage-0\") pod \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.830330 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76x4g\" (UniqueName: \"kubernetes.io/projected/8e8eb649-79cf-41a7-a2be-a0d713c6d741-kube-api-access-76x4g\") pod \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\" (UID: \"8e8eb649-79cf-41a7-a2be-a0d713c6d741\") " Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.841141 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e8eb649-79cf-41a7-a2be-a0d713c6d741-kube-api-access-76x4g" (OuterVolumeSpecName: "kube-api-access-76x4g") pod "8e8eb649-79cf-41a7-a2be-a0d713c6d741" (UID: "8e8eb649-79cf-41a7-a2be-a0d713c6d741"). InnerVolumeSpecName "kube-api-access-76x4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.920948 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e8eb649-79cf-41a7-a2be-a0d713c6d741" (UID: "8e8eb649-79cf-41a7-a2be-a0d713c6d741"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.932018 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e8eb649-79cf-41a7-a2be-a0d713c6d741" (UID: "8e8eb649-79cf-41a7-a2be-a0d713c6d741"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.933644 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.933710 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.933723 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76x4g\" (UniqueName: \"kubernetes.io/projected/8e8eb649-79cf-41a7-a2be-a0d713c6d741-kube-api-access-76x4g\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.936873 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-config" (OuterVolumeSpecName: "config") pod "8e8eb649-79cf-41a7-a2be-a0d713c6d741" (UID: "8e8eb649-79cf-41a7-a2be-a0d713c6d741"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.941841 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e8eb649-79cf-41a7-a2be-a0d713c6d741" (UID: "8e8eb649-79cf-41a7-a2be-a0d713c6d741"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:26 crc kubenswrapper[4992]: I1004 04:54:26.974063 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8e8eb649-79cf-41a7-a2be-a0d713c6d741" (UID: "8e8eb649-79cf-41a7-a2be-a0d713c6d741"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.035375 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.035669 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.035679 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8eb649-79cf-41a7-a2be-a0d713c6d741-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.068129 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 04:54:27 crc kubenswrapper[4992]: W1004 04:54:27.068254 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a864935_07eb_48c7_a02b_a81ff050ce4b.slice/crio-9a8c7d91f20156932c52c2254857ad9db5a82ee62b7af4d621b3165dfc37d942 WatchSource:0}: Error finding container 9a8c7d91f20156932c52c2254857ad9db5a82ee62b7af4d621b3165dfc37d942: Status 404 returned error can't find the container with id 9a8c7d91f20156932c52c2254857ad9db5a82ee62b7af4d621b3165dfc37d942 Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.463806 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" event={"ID":"8e8eb649-79cf-41a7-a2be-a0d713c6d741","Type":"ContainerDied","Data":"ff4ae11e26c176575c66a03fbaeccd29d82a31185804d7a38f920b41b643aaa2"} Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.464163 4992 scope.go:117] "RemoveContainer" containerID="48780c5b8756483e0bceadf4f9d0c6c2d167eade4004c20641ca7c79af35cbb5" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.463847 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-qs2sd" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.464989 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8a864935-07eb-48c7-a02b-a81ff050ce4b","Type":"ContainerStarted","Data":"9a8c7d91f20156932c52c2254857ad9db5a82ee62b7af4d621b3165dfc37d942"} Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.470417 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6d24bc2b-9416-4bf0-aff1-081826b9c9d9","Type":"ContainerStarted","Data":"801c98346a059414237a17e9e68f1b7a3cddabb7dba81799ab7a7efc170b839b"} Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.470463 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.488346 4992 scope.go:117] "RemoveContainer" containerID="b51c0840455bb7516008bf91067006436cbf57cc556eeedf245977b73854a864" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.500599 4992 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="f5fac4a9-5b82-4908-a798-69469297f7e6" podUID="8a864935-07eb-48c7-a02b-a81ff050ce4b" Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.509911 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-qs2sd"] Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.516816 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-qs2sd"] Oct 04 04:54:27 crc kubenswrapper[4992]: I1004 04:54:27.775588 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 04:54:28 crc kubenswrapper[4992]: I1004 04:54:28.348260 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" path="/var/lib/kubelet/pods/8e8eb649-79cf-41a7-a2be-a0d713c6d741/volumes" Oct 04 04:54:28 crc kubenswrapper[4992]: I1004 04:54:28.349468 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5fac4a9-5b82-4908-a798-69469297f7e6" path="/var/lib/kubelet/pods/f5fac4a9-5b82-4908-a798-69469297f7e6/volumes" Oct 04 04:54:28 crc kubenswrapper[4992]: I1004 04:54:28.508032 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.508017192 podStartE2EDuration="4.508017192s" podCreationTimestamp="2025-10-04 04:54:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:28.500273256 +0000 UTC m=+1302.347948734" watchObservedRunningTime="2025-10-04 04:54:28.508017192 +0000 UTC m=+1302.355692660" Oct 04 04:54:29 crc kubenswrapper[4992]: I1004 04:54:29.801583 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.677588 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-597c8b5b9f-s2g6d"] Oct 04 04:54:31 crc kubenswrapper[4992]: E1004 04:54:31.678452 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerName="dnsmasq-dns" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.678466 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerName="dnsmasq-dns" Oct 04 04:54:31 crc kubenswrapper[4992]: E1004 04:54:31.678485 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerName="init" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.678491 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerName="init" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.678710 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8eb649-79cf-41a7-a2be-a0d713c6d741" containerName="dnsmasq-dns" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.679757 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.689089 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.689172 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.690938 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.696811 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-597c8b5b9f-s2g6d"] Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.722315 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mf5c\" (UniqueName: \"kubernetes.io/projected/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-kube-api-access-7mf5c\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.722382 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-combined-ca-bundle\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.722426 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-config-data\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.722457 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-log-httpd\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.722497 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-etc-swift\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.722526 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-run-httpd\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.722541 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-public-tls-certs\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.722614 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-internal-tls-certs\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.823848 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-etc-swift\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.823921 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-run-httpd\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.823953 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-public-tls-certs\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.824057 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-internal-tls-certs\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.824088 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mf5c\" (UniqueName: \"kubernetes.io/projected/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-kube-api-access-7mf5c\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.824114 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-combined-ca-bundle\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.824162 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-config-data\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.824201 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-log-httpd\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.824884 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-log-httpd\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.826176 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-run-httpd\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.832110 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-internal-tls-certs\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.832672 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-combined-ca-bundle\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.836865 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-etc-swift\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.838950 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-config-data\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.845263 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mf5c\" (UniqueName: \"kubernetes.io/projected/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-kube-api-access-7mf5c\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:31 crc kubenswrapper[4992]: I1004 04:54:31.858609 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab81341f-2395-4a8d-a6c0-04ad2c2f970a-public-tls-certs\") pod \"swift-proxy-597c8b5b9f-s2g6d\" (UID: \"ab81341f-2395-4a8d-a6c0-04ad2c2f970a\") " pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:32 crc kubenswrapper[4992]: I1004 04:54:32.002390 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:32 crc kubenswrapper[4992]: I1004 04:54:32.536820 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:32 crc kubenswrapper[4992]: I1004 04:54:32.538669 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="proxy-httpd" containerID="cri-o://076434825bfa10ede425a2c4520f91bdb8f34d64169e9ae4b348d391bd9eb0db" gracePeriod=30 Oct 04 04:54:32 crc kubenswrapper[4992]: I1004 04:54:32.538796 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="sg-core" containerID="cri-o://3e8fdcdab7c76f4ad8759bf84326581df5b6b4b5687ae287afffa25cae3e52eb" gracePeriod=30 Oct 04 04:54:32 crc kubenswrapper[4992]: I1004 04:54:32.538868 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="ceilometer-notification-agent" containerID="cri-o://44965bdbac199eba42f3c8cc171aeec494582286882e43a069cecbe2b513faa8" gracePeriod=30 Oct 04 04:54:32 crc kubenswrapper[4992]: I1004 04:54:32.539130 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="ceilometer-central-agent" containerID="cri-o://9ae04bc1b53e75d2f170ef9ccb3a4ea7ed93f4874c42f94844a1a1fb2c2b303b" gracePeriod=30 Oct 04 04:54:32 crc kubenswrapper[4992]: I1004 04:54:32.546776 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.161:3000/\": EOF" Oct 04 04:54:33 crc kubenswrapper[4992]: I1004 04:54:33.537120 4992 generic.go:334] "Generic (PLEG): container finished" podID="933782b7-db86-44ca-97e9-34ae542598a4" containerID="076434825bfa10ede425a2c4520f91bdb8f34d64169e9ae4b348d391bd9eb0db" exitCode=0 Oct 04 04:54:33 crc kubenswrapper[4992]: I1004 04:54:33.537484 4992 generic.go:334] "Generic (PLEG): container finished" podID="933782b7-db86-44ca-97e9-34ae542598a4" containerID="3e8fdcdab7c76f4ad8759bf84326581df5b6b4b5687ae287afffa25cae3e52eb" exitCode=2 Oct 04 04:54:33 crc kubenswrapper[4992]: I1004 04:54:33.537500 4992 generic.go:334] "Generic (PLEG): container finished" podID="933782b7-db86-44ca-97e9-34ae542598a4" containerID="9ae04bc1b53e75d2f170ef9ccb3a4ea7ed93f4874c42f94844a1a1fb2c2b303b" exitCode=0 Oct 04 04:54:33 crc kubenswrapper[4992]: I1004 04:54:33.537229 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerDied","Data":"076434825bfa10ede425a2c4520f91bdb8f34d64169e9ae4b348d391bd9eb0db"} Oct 04 04:54:33 crc kubenswrapper[4992]: I1004 04:54:33.537541 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerDied","Data":"3e8fdcdab7c76f4ad8759bf84326581df5b6b4b5687ae287afffa25cae3e52eb"} Oct 04 04:54:33 crc kubenswrapper[4992]: I1004 04:54:33.537558 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerDied","Data":"9ae04bc1b53e75d2f170ef9ccb3a4ea7ed93f4874c42f94844a1a1fb2c2b303b"} Oct 04 04:54:35 crc kubenswrapper[4992]: I1004 04:54:35.035889 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 04:54:36 crc kubenswrapper[4992]: I1004 04:54:36.182320 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:36 crc kubenswrapper[4992]: I1004 04:54:36.208149 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-689c5fd8b4-p4lxm" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.311934 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-h7lnf"] Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.313719 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-h7lnf" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.346055 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-h7lnf"] Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.418908 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-5pxw8"] Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.420419 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5pxw8" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.428327 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-5pxw8"] Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.439636 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4p2q\" (UniqueName: \"kubernetes.io/projected/c3343348-309c-4770-8459-977ee71a5571-kube-api-access-v4p2q\") pod \"nova-api-db-create-h7lnf\" (UID: \"c3343348-309c-4770-8459-977ee71a5571\") " pod="openstack/nova-api-db-create-h7lnf" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.514999 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-62css"] Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.516708 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-62css" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.523633 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-62css"] Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.541852 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4p2q\" (UniqueName: \"kubernetes.io/projected/c3343348-309c-4770-8459-977ee71a5571-kube-api-access-v4p2q\") pod \"nova-api-db-create-h7lnf\" (UID: \"c3343348-309c-4770-8459-977ee71a5571\") " pod="openstack/nova-api-db-create-h7lnf" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.541954 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c42zx\" (UniqueName: \"kubernetes.io/projected/e3b4bd43-8554-475a-9cc6-af381cc236e4-kube-api-access-c42zx\") pod \"nova-cell0-db-create-5pxw8\" (UID: \"e3b4bd43-8554-475a-9cc6-af381cc236e4\") " pod="openstack/nova-cell0-db-create-5pxw8" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.569657 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4p2q\" (UniqueName: \"kubernetes.io/projected/c3343348-309c-4770-8459-977ee71a5571-kube-api-access-v4p2q\") pod \"nova-api-db-create-h7lnf\" (UID: \"c3343348-309c-4770-8459-977ee71a5571\") " pod="openstack/nova-api-db-create-h7lnf" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.643832 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c42zx\" (UniqueName: \"kubernetes.io/projected/e3b4bd43-8554-475a-9cc6-af381cc236e4-kube-api-access-c42zx\") pod \"nova-cell0-db-create-5pxw8\" (UID: \"e3b4bd43-8554-475a-9cc6-af381cc236e4\") " pod="openstack/nova-cell0-db-create-5pxw8" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.644283 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24zkd\" (UniqueName: \"kubernetes.io/projected/c5340c92-9a3c-42f7-abbe-1f5e1cd2572b-kube-api-access-24zkd\") pod \"nova-cell1-db-create-62css\" (UID: \"c5340c92-9a3c-42f7-abbe-1f5e1cd2572b\") " pod="openstack/nova-cell1-db-create-62css" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.665432 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c42zx\" (UniqueName: \"kubernetes.io/projected/e3b4bd43-8554-475a-9cc6-af381cc236e4-kube-api-access-c42zx\") pod \"nova-cell0-db-create-5pxw8\" (UID: \"e3b4bd43-8554-475a-9cc6-af381cc236e4\") " pod="openstack/nova-cell0-db-create-5pxw8" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.746226 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24zkd\" (UniqueName: \"kubernetes.io/projected/c5340c92-9a3c-42f7-abbe-1f5e1cd2572b-kube-api-access-24zkd\") pod \"nova-cell1-db-create-62css\" (UID: \"c5340c92-9a3c-42f7-abbe-1f5e1cd2572b\") " pod="openstack/nova-cell1-db-create-62css" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.747816 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-h7lnf" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.753257 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5pxw8" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.763146 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24zkd\" (UniqueName: \"kubernetes.io/projected/c5340c92-9a3c-42f7-abbe-1f5e1cd2572b-kube-api-access-24zkd\") pod \"nova-cell1-db-create-62css\" (UID: \"c5340c92-9a3c-42f7-abbe-1f5e1cd2572b\") " pod="openstack/nova-cell1-db-create-62css" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.878640 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-62css" Oct 04 04:54:37 crc kubenswrapper[4992]: I1004 04:54:37.901958 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-597c8b5b9f-s2g6d"] Oct 04 04:54:37 crc kubenswrapper[4992]: W1004 04:54:37.936017 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab81341f_2395_4a8d_a6c0_04ad2c2f970a.slice/crio-a992bedef7380a84eaa23a6247cb1ca224a77cddb4f213bab71c0822ad056d92 WatchSource:0}: Error finding container a992bedef7380a84eaa23a6247cb1ca224a77cddb4f213bab71c0822ad056d92: Status 404 returned error can't find the container with id a992bedef7380a84eaa23a6247cb1ca224a77cddb4f213bab71c0822ad056d92 Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.300703 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-h7lnf"] Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.399909 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-5pxw8"] Oct 04 04:54:38 crc kubenswrapper[4992]: W1004 04:54:38.414934 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3b4bd43_8554_475a_9cc6_af381cc236e4.slice/crio-82b9b80b2a76dd2e8e31a335539168f65895d3f1edf0d604ef9abae7eb63e7c4 WatchSource:0}: Error finding container 82b9b80b2a76dd2e8e31a335539168f65895d3f1edf0d604ef9abae7eb63e7c4: Status 404 returned error can't find the container with id 82b9b80b2a76dd2e8e31a335539168f65895d3f1edf0d604ef9abae7eb63e7c4 Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.504983 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-62css"] Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.602639 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-62css" event={"ID":"c5340c92-9a3c-42f7-abbe-1f5e1cd2572b","Type":"ContainerStarted","Data":"0f9384d9af0a8ff76eba662f19eac1e705ec44a1711a87c8b3fdf7ef314a6661"} Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.606142 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5pxw8" event={"ID":"e3b4bd43-8554-475a-9cc6-af381cc236e4","Type":"ContainerStarted","Data":"82b9b80b2a76dd2e8e31a335539168f65895d3f1edf0d604ef9abae7eb63e7c4"} Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.610163 4992 generic.go:334] "Generic (PLEG): container finished" podID="933782b7-db86-44ca-97e9-34ae542598a4" containerID="44965bdbac199eba42f3c8cc171aeec494582286882e43a069cecbe2b513faa8" exitCode=0 Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.610223 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerDied","Data":"44965bdbac199eba42f3c8cc171aeec494582286882e43a069cecbe2b513faa8"} Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.613860 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8a864935-07eb-48c7-a02b-a81ff050ce4b","Type":"ContainerStarted","Data":"96c4bd263ec81b0fc47ec38e7f919be4040ebd906b7b0c34d634945b5e5b0de9"} Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.617618 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-h7lnf" event={"ID":"c3343348-309c-4770-8459-977ee71a5571","Type":"ContainerStarted","Data":"3e6e05f53a8ad5313a292a23308dfab3639f9064ce904c9f65463910deec1e78"} Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.619092 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" event={"ID":"ab81341f-2395-4a8d-a6c0-04ad2c2f970a","Type":"ContainerStarted","Data":"5525de81a6182332a11629ff345ac00519492bb5f4753454977f1942f818c7d2"} Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.619114 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" event={"ID":"ab81341f-2395-4a8d-a6c0-04ad2c2f970a","Type":"ContainerStarted","Data":"a992bedef7380a84eaa23a6247cb1ca224a77cddb4f213bab71c0822ad056d92"} Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.629858 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.434799206 podStartE2EDuration="12.629840187s" podCreationTimestamp="2025-10-04 04:54:26 +0000 UTC" firstStartedPulling="2025-10-04 04:54:27.075119013 +0000 UTC m=+1300.922794481" lastFinishedPulling="2025-10-04 04:54:37.270159994 +0000 UTC m=+1311.117835462" observedRunningTime="2025-10-04 04:54:38.627160265 +0000 UTC m=+1312.474835733" watchObservedRunningTime="2025-10-04 04:54:38.629840187 +0000 UTC m=+1312.477515655" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.647971 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.780572 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-sg-core-conf-yaml\") pod \"933782b7-db86-44ca-97e9-34ae542598a4\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.780651 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-combined-ca-bundle\") pod \"933782b7-db86-44ca-97e9-34ae542598a4\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.780807 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-log-httpd\") pod \"933782b7-db86-44ca-97e9-34ae542598a4\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.780853 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rdk7\" (UniqueName: \"kubernetes.io/projected/933782b7-db86-44ca-97e9-34ae542598a4-kube-api-access-2rdk7\") pod \"933782b7-db86-44ca-97e9-34ae542598a4\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.780884 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-config-data\") pod \"933782b7-db86-44ca-97e9-34ae542598a4\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.780952 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-scripts\") pod \"933782b7-db86-44ca-97e9-34ae542598a4\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.780981 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-run-httpd\") pod \"933782b7-db86-44ca-97e9-34ae542598a4\" (UID: \"933782b7-db86-44ca-97e9-34ae542598a4\") " Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.781798 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "933782b7-db86-44ca-97e9-34ae542598a4" (UID: "933782b7-db86-44ca-97e9-34ae542598a4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.782796 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "933782b7-db86-44ca-97e9-34ae542598a4" (UID: "933782b7-db86-44ca-97e9-34ae542598a4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.787690 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-scripts" (OuterVolumeSpecName: "scripts") pod "933782b7-db86-44ca-97e9-34ae542598a4" (UID: "933782b7-db86-44ca-97e9-34ae542598a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.792892 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/933782b7-db86-44ca-97e9-34ae542598a4-kube-api-access-2rdk7" (OuterVolumeSpecName: "kube-api-access-2rdk7") pod "933782b7-db86-44ca-97e9-34ae542598a4" (UID: "933782b7-db86-44ca-97e9-34ae542598a4"). InnerVolumeSpecName "kube-api-access-2rdk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.809217 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "933782b7-db86-44ca-97e9-34ae542598a4" (UID: "933782b7-db86-44ca-97e9-34ae542598a4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.876826 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "933782b7-db86-44ca-97e9-34ae542598a4" (UID: "933782b7-db86-44ca-97e9-34ae542598a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.883549 4992 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.883677 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rdk7\" (UniqueName: \"kubernetes.io/projected/933782b7-db86-44ca-97e9-34ae542598a4-kube-api-access-2rdk7\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.883691 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.883705 4992 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/933782b7-db86-44ca-97e9-34ae542598a4-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.883716 4992 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.883727 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.898582 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-config-data" (OuterVolumeSpecName: "config-data") pod "933782b7-db86-44ca-97e9-34ae542598a4" (UID: "933782b7-db86-44ca-97e9-34ae542598a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:38 crc kubenswrapper[4992]: I1004 04:54:38.988234 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933782b7-db86-44ca-97e9-34ae542598a4-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.630994 4992 generic.go:334] "Generic (PLEG): container finished" podID="c5340c92-9a3c-42f7-abbe-1f5e1cd2572b" containerID="6d064c4627574d8dfe5a62ac0c60602fbca42898f81558091d71052804a329e4" exitCode=0 Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.631057 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-62css" event={"ID":"c5340c92-9a3c-42f7-abbe-1f5e1cd2572b","Type":"ContainerDied","Data":"6d064c4627574d8dfe5a62ac0c60602fbca42898f81558091d71052804a329e4"} Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.636148 4992 generic.go:334] "Generic (PLEG): container finished" podID="e3b4bd43-8554-475a-9cc6-af381cc236e4" containerID="2019828bb9888d08b01787093f8c88dccc9f6fd7ebc29f21a6c644942b216844" exitCode=0 Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.636213 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5pxw8" event={"ID":"e3b4bd43-8554-475a-9cc6-af381cc236e4","Type":"ContainerDied","Data":"2019828bb9888d08b01787093f8c88dccc9f6fd7ebc29f21a6c644942b216844"} Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.643077 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"933782b7-db86-44ca-97e9-34ae542598a4","Type":"ContainerDied","Data":"18e78b34097a5cdd3f444f2ee1c3d88525af6b6b173a3f87202346f36b082d02"} Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.643125 4992 scope.go:117] "RemoveContainer" containerID="076434825bfa10ede425a2c4520f91bdb8f34d64169e9ae4b348d391bd9eb0db" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.643259 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.654626 4992 generic.go:334] "Generic (PLEG): container finished" podID="c3343348-309c-4770-8459-977ee71a5571" containerID="e288880fc1464923ab57c5c368f1b090af95a43477c93f6c31a2e4b80f40ac42" exitCode=0 Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.654713 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-h7lnf" event={"ID":"c3343348-309c-4770-8459-977ee71a5571","Type":"ContainerDied","Data":"e288880fc1464923ab57c5c368f1b090af95a43477c93f6c31a2e4b80f40ac42"} Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.663112 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" event={"ID":"ab81341f-2395-4a8d-a6c0-04ad2c2f970a","Type":"ContainerStarted","Data":"fa10ca8e50bdf512cc11ec17ec6c15cc192635b4ede802a16c4d6bd084a598ab"} Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.663255 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.663270 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.695736 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.702897 4992 scope.go:117] "RemoveContainer" containerID="3e8fdcdab7c76f4ad8759bf84326581df5b6b4b5687ae287afffa25cae3e52eb" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.711479 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.736724 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:39 crc kubenswrapper[4992]: E1004 04:54:39.737186 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="ceilometer-central-agent" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.737206 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="ceilometer-central-agent" Oct 04 04:54:39 crc kubenswrapper[4992]: E1004 04:54:39.737222 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="sg-core" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.737228 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="sg-core" Oct 04 04:54:39 crc kubenswrapper[4992]: E1004 04:54:39.737242 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="proxy-httpd" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.737248 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="proxy-httpd" Oct 04 04:54:39 crc kubenswrapper[4992]: E1004 04:54:39.737258 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="ceilometer-notification-agent" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.737263 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="ceilometer-notification-agent" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.737455 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="sg-core" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.737478 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="ceilometer-central-agent" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.737496 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="ceilometer-notification-agent" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.737507 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="933782b7-db86-44ca-97e9-34ae542598a4" containerName="proxy-httpd" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.739387 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.739545 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.740138 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" podStartSLOduration=8.7401168 podStartE2EDuration="8.7401168s" podCreationTimestamp="2025-10-04 04:54:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:54:39.738734863 +0000 UTC m=+1313.586410331" watchObservedRunningTime="2025-10-04 04:54:39.7401168 +0000 UTC m=+1313.587792268" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.750338 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.750591 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.774746 4992 scope.go:117] "RemoveContainer" containerID="44965bdbac199eba42f3c8cc171aeec494582286882e43a069cecbe2b513faa8" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.806677 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-log-httpd\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.806719 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-config-data\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.806739 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfclx\" (UniqueName: \"kubernetes.io/projected/bbafcdf6-a834-4c88-9c0f-6af2dec470db-kube-api-access-sfclx\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.806755 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-scripts\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.806816 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.806983 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.807009 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-run-httpd\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.853119 4992 scope.go:117] "RemoveContainer" containerID="9ae04bc1b53e75d2f170ef9ccb3a4ea7ed93f4874c42f94844a1a1fb2c2b303b" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.909207 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.909255 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-run-httpd\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.909302 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-log-httpd\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.909318 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-config-data\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.909337 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfclx\" (UniqueName: \"kubernetes.io/projected/bbafcdf6-a834-4c88-9c0f-6af2dec470db-kube-api-access-sfclx\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.909354 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-scripts\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.909420 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.911061 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-log-httpd\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.916809 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-run-httpd\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.925167 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-config-data\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.928953 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.948918 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-scripts\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.950657 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:39 crc kubenswrapper[4992]: I1004 04:54:39.951926 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfclx\" (UniqueName: \"kubernetes.io/projected/bbafcdf6-a834-4c88-9c0f-6af2dec470db-kube-api-access-sfclx\") pod \"ceilometer-0\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " pod="openstack/ceilometer-0" Oct 04 04:54:40 crc kubenswrapper[4992]: I1004 04:54:40.106209 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:40 crc kubenswrapper[4992]: I1004 04:54:40.336392 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="933782b7-db86-44ca-97e9-34ae542598a4" path="/var/lib/kubelet/pods/933782b7-db86-44ca-97e9-34ae542598a4/volumes" Oct 04 04:54:40 crc kubenswrapper[4992]: I1004 04:54:40.648436 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:40 crc kubenswrapper[4992]: W1004 04:54:40.654803 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbafcdf6_a834_4c88_9c0f_6af2dec470db.slice/crio-4d17c4488d400c3d34c4d481e9e7f495b4e0813ee481ecd3bb8a61c5946dbef3 WatchSource:0}: Error finding container 4d17c4488d400c3d34c4d481e9e7f495b4e0813ee481ecd3bb8a61c5946dbef3: Status 404 returned error can't find the container with id 4d17c4488d400c3d34c4d481e9e7f495b4e0813ee481ecd3bb8a61c5946dbef3 Oct 04 04:54:40 crc kubenswrapper[4992]: I1004 04:54:40.679968 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerStarted","Data":"4d17c4488d400c3d34c4d481e9e7f495b4e0813ee481ecd3bb8a61c5946dbef3"} Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.150968 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-h7lnf" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.239286 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4p2q\" (UniqueName: \"kubernetes.io/projected/c3343348-309c-4770-8459-977ee71a5571-kube-api-access-v4p2q\") pod \"c3343348-309c-4770-8459-977ee71a5571\" (UID: \"c3343348-309c-4770-8459-977ee71a5571\") " Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.243033 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3343348-309c-4770-8459-977ee71a5571-kube-api-access-v4p2q" (OuterVolumeSpecName: "kube-api-access-v4p2q") pod "c3343348-309c-4770-8459-977ee71a5571" (UID: "c3343348-309c-4770-8459-977ee71a5571"). InnerVolumeSpecName "kube-api-access-v4p2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.324375 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.329466 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5pxw8" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.335147 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-62css" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.341109 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4p2q\" (UniqueName: \"kubernetes.io/projected/c3343348-309c-4770-8459-977ee71a5571-kube-api-access-v4p2q\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.442289 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c42zx\" (UniqueName: \"kubernetes.io/projected/e3b4bd43-8554-475a-9cc6-af381cc236e4-kube-api-access-c42zx\") pod \"e3b4bd43-8554-475a-9cc6-af381cc236e4\" (UID: \"e3b4bd43-8554-475a-9cc6-af381cc236e4\") " Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.442514 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24zkd\" (UniqueName: \"kubernetes.io/projected/c5340c92-9a3c-42f7-abbe-1f5e1cd2572b-kube-api-access-24zkd\") pod \"c5340c92-9a3c-42f7-abbe-1f5e1cd2572b\" (UID: \"c5340c92-9a3c-42f7-abbe-1f5e1cd2572b\") " Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.453747 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3b4bd43-8554-475a-9cc6-af381cc236e4-kube-api-access-c42zx" (OuterVolumeSpecName: "kube-api-access-c42zx") pod "e3b4bd43-8554-475a-9cc6-af381cc236e4" (UID: "e3b4bd43-8554-475a-9cc6-af381cc236e4"). InnerVolumeSpecName "kube-api-access-c42zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.453909 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5340c92-9a3c-42f7-abbe-1f5e1cd2572b-kube-api-access-24zkd" (OuterVolumeSpecName: "kube-api-access-24zkd") pod "c5340c92-9a3c-42f7-abbe-1f5e1cd2572b" (UID: "c5340c92-9a3c-42f7-abbe-1f5e1cd2572b"). InnerVolumeSpecName "kube-api-access-24zkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.544255 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24zkd\" (UniqueName: \"kubernetes.io/projected/c5340c92-9a3c-42f7-abbe-1f5e1cd2572b-kube-api-access-24zkd\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.544286 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c42zx\" (UniqueName: \"kubernetes.io/projected/e3b4bd43-8554-475a-9cc6-af381cc236e4-kube-api-access-c42zx\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.692740 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-h7lnf" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.692739 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-h7lnf" event={"ID":"c3343348-309c-4770-8459-977ee71a5571","Type":"ContainerDied","Data":"3e6e05f53a8ad5313a292a23308dfab3639f9064ce904c9f65463910deec1e78"} Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.692865 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e6e05f53a8ad5313a292a23308dfab3639f9064ce904c9f65463910deec1e78" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.703645 4992 generic.go:334] "Generic (PLEG): container finished" podID="94c198d3-0808-4077-951a-fe14a94409ea" containerID="5360ba4db45400a267ecaceaa1c0924dabeb2f6d586b197d48d547360639404f" exitCode=137 Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.703718 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6dc74865-2rgnq" event={"ID":"94c198d3-0808-4077-951a-fe14a94409ea","Type":"ContainerDied","Data":"5360ba4db45400a267ecaceaa1c0924dabeb2f6d586b197d48d547360639404f"} Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.708643 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerStarted","Data":"943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754"} Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.714542 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-62css" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.714531 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-62css" event={"ID":"c5340c92-9a3c-42f7-abbe-1f5e1cd2572b","Type":"ContainerDied","Data":"0f9384d9af0a8ff76eba662f19eac1e705ec44a1711a87c8b3fdf7ef314a6661"} Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.714693 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f9384d9af0a8ff76eba662f19eac1e705ec44a1711a87c8b3fdf7ef314a6661" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.722180 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5pxw8" event={"ID":"e3b4bd43-8554-475a-9cc6-af381cc236e4","Type":"ContainerDied","Data":"82b9b80b2a76dd2e8e31a335539168f65895d3f1edf0d604ef9abae7eb63e7c4"} Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.722228 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82b9b80b2a76dd2e8e31a335539168f65895d3f1edf0d604ef9abae7eb63e7c4" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.722244 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5pxw8" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.874641 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.950874 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c198d3-0808-4077-951a-fe14a94409ea-logs\") pod \"94c198d3-0808-4077-951a-fe14a94409ea\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.950994 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94c198d3-0808-4077-951a-fe14a94409ea-horizon-secret-key\") pod \"94c198d3-0808-4077-951a-fe14a94409ea\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.951082 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-scripts\") pod \"94c198d3-0808-4077-951a-fe14a94409ea\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.951128 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6f8c\" (UniqueName: \"kubernetes.io/projected/94c198d3-0808-4077-951a-fe14a94409ea-kube-api-access-g6f8c\") pod \"94c198d3-0808-4077-951a-fe14a94409ea\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.951200 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-config-data\") pod \"94c198d3-0808-4077-951a-fe14a94409ea\" (UID: \"94c198d3-0808-4077-951a-fe14a94409ea\") " Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.951347 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94c198d3-0808-4077-951a-fe14a94409ea-logs" (OuterVolumeSpecName: "logs") pod "94c198d3-0808-4077-951a-fe14a94409ea" (UID: "94c198d3-0808-4077-951a-fe14a94409ea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.951690 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94c198d3-0808-4077-951a-fe14a94409ea-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.954849 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94c198d3-0808-4077-951a-fe14a94409ea-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "94c198d3-0808-4077-951a-fe14a94409ea" (UID: "94c198d3-0808-4077-951a-fe14a94409ea"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.955638 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94c198d3-0808-4077-951a-fe14a94409ea-kube-api-access-g6f8c" (OuterVolumeSpecName: "kube-api-access-g6f8c") pod "94c198d3-0808-4077-951a-fe14a94409ea" (UID: "94c198d3-0808-4077-951a-fe14a94409ea"). InnerVolumeSpecName "kube-api-access-g6f8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.974906 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-scripts" (OuterVolumeSpecName: "scripts") pod "94c198d3-0808-4077-951a-fe14a94409ea" (UID: "94c198d3-0808-4077-951a-fe14a94409ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:41 crc kubenswrapper[4992]: I1004 04:54:41.975255 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-config-data" (OuterVolumeSpecName: "config-data") pod "94c198d3-0808-4077-951a-fe14a94409ea" (UID: "94c198d3-0808-4077-951a-fe14a94409ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.054292 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.054869 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6f8c\" (UniqueName: \"kubernetes.io/projected/94c198d3-0808-4077-951a-fe14a94409ea-kube-api-access-g6f8c\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.054890 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94c198d3-0808-4077-951a-fe14a94409ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.054903 4992 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/94c198d3-0808-4077-951a-fe14a94409ea-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.745513 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerStarted","Data":"0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6"} Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.749092 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6dc74865-2rgnq" event={"ID":"94c198d3-0808-4077-951a-fe14a94409ea","Type":"ContainerDied","Data":"2523dae3be6d8dd6670c2ec26687a5726194a8d297a671c1803af2a5938e33f6"} Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.749148 4992 scope.go:117] "RemoveContainer" containerID="0e838cf84f8b2286de7fee1771b85b59290fdf40b14d9269125d5379345a8111" Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.749214 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6dc74865-2rgnq" Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.910117 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c6dc74865-2rgnq"] Oct 04 04:54:42 crc kubenswrapper[4992]: I1004 04:54:42.922830 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c6dc74865-2rgnq"] Oct 04 04:54:43 crc kubenswrapper[4992]: I1004 04:54:43.020393 4992 scope.go:117] "RemoveContainer" containerID="5360ba4db45400a267ecaceaa1c0924dabeb2f6d586b197d48d547360639404f" Oct 04 04:54:43 crc kubenswrapper[4992]: I1004 04:54:43.758720 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerStarted","Data":"4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95"} Oct 04 04:54:44 crc kubenswrapper[4992]: I1004 04:54:44.341106 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94c198d3-0808-4077-951a-fe14a94409ea" path="/var/lib/kubelet/pods/94c198d3-0808-4077-951a-fe14a94409ea/volumes" Oct 04 04:54:44 crc kubenswrapper[4992]: I1004 04:54:44.774587 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerStarted","Data":"3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187"} Oct 04 04:54:44 crc kubenswrapper[4992]: I1004 04:54:44.775301 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="ceilometer-central-agent" containerID="cri-o://943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754" gracePeriod=30 Oct 04 04:54:44 crc kubenswrapper[4992]: I1004 04:54:44.775899 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:54:44 crc kubenswrapper[4992]: I1004 04:54:44.775397 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="proxy-httpd" containerID="cri-o://3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187" gracePeriod=30 Oct 04 04:54:44 crc kubenswrapper[4992]: I1004 04:54:44.775377 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="sg-core" containerID="cri-o://4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95" gracePeriod=30 Oct 04 04:54:44 crc kubenswrapper[4992]: I1004 04:54:44.775411 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="ceilometer-notification-agent" containerID="cri-o://0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6" gracePeriod=30 Oct 04 04:54:44 crc kubenswrapper[4992]: I1004 04:54:44.805576 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.175024041 podStartE2EDuration="5.805556997s" podCreationTimestamp="2025-10-04 04:54:39 +0000 UTC" firstStartedPulling="2025-10-04 04:54:40.656609621 +0000 UTC m=+1314.504285089" lastFinishedPulling="2025-10-04 04:54:44.287142577 +0000 UTC m=+1318.134818045" observedRunningTime="2025-10-04 04:54:44.801215382 +0000 UTC m=+1318.648890870" watchObservedRunningTime="2025-10-04 04:54:44.805556997 +0000 UTC m=+1318.653232465" Oct 04 04:54:45 crc kubenswrapper[4992]: I1004 04:54:45.793727 4992 generic.go:334] "Generic (PLEG): container finished" podID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerID="3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187" exitCode=0 Oct 04 04:54:45 crc kubenswrapper[4992]: I1004 04:54:45.793762 4992 generic.go:334] "Generic (PLEG): container finished" podID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerID="4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95" exitCode=2 Oct 04 04:54:45 crc kubenswrapper[4992]: I1004 04:54:45.793781 4992 generic.go:334] "Generic (PLEG): container finished" podID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerID="0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6" exitCode=0 Oct 04 04:54:45 crc kubenswrapper[4992]: I1004 04:54:45.793804 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerDied","Data":"3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187"} Oct 04 04:54:45 crc kubenswrapper[4992]: I1004 04:54:45.793831 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerDied","Data":"4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95"} Oct 04 04:54:45 crc kubenswrapper[4992]: I1004 04:54:45.793841 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerDied","Data":"0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6"} Oct 04 04:54:45 crc kubenswrapper[4992]: I1004 04:54:45.944756 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.028032 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.030053 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-597c8b5b9f-s2g6d" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.228977 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.364783 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-663d-account-create-vr8wb"] Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365216 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3b4bd43-8554-475a-9cc6-af381cc236e4" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365233 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3b4bd43-8554-475a-9cc6-af381cc236e4" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365252 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365259 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365271 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon-log" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365279 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon-log" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365299 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5340c92-9a3c-42f7-abbe-1f5e1cd2572b" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365309 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5340c92-9a3c-42f7-abbe-1f5e1cd2572b" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365319 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="proxy-httpd" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365324 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="proxy-httpd" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365333 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="sg-core" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365338 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="sg-core" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365347 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="ceilometer-notification-agent" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365370 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="ceilometer-notification-agent" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365388 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3343348-309c-4770-8459-977ee71a5571" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365394 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3343348-309c-4770-8459-977ee71a5571" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.365407 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="ceilometer-central-agent" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365413 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="ceilometer-central-agent" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365569 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3343348-309c-4770-8459-977ee71a5571" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365578 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="ceilometer-notification-agent" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365589 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon-log" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365603 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5340c92-9a3c-42f7-abbe-1f5e1cd2572b" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365617 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3b4bd43-8554-475a-9cc6-af381cc236e4" containerName="mariadb-database-create" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365627 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="ceilometer-central-agent" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365634 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="94c198d3-0808-4077-951a-fe14a94409ea" containerName="horizon" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365650 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="sg-core" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.365660 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerName="proxy-httpd" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.366413 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-663d-account-create-vr8wb" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.366724 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-combined-ca-bundle\") pod \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.366760 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-scripts\") pod \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.366837 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-log-httpd\") pod \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.366923 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-config-data\") pod \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.366972 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfclx\" (UniqueName: \"kubernetes.io/projected/bbafcdf6-a834-4c88-9c0f-6af2dec470db-kube-api-access-sfclx\") pod \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.367040 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-sg-core-conf-yaml\") pod \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.367062 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-run-httpd\") pod \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\" (UID: \"bbafcdf6-a834-4c88-9c0f-6af2dec470db\") " Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.367638 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbafcdf6-a834-4c88-9c0f-6af2dec470db" (UID: "bbafcdf6-a834-4c88-9c0f-6af2dec470db"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.367660 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbafcdf6-a834-4c88-9c0f-6af2dec470db" (UID: "bbafcdf6-a834-4c88-9c0f-6af2dec470db"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.368753 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.373915 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-scripts" (OuterVolumeSpecName: "scripts") pod "bbafcdf6-a834-4c88-9c0f-6af2dec470db" (UID: "bbafcdf6-a834-4c88-9c0f-6af2dec470db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.374430 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbafcdf6-a834-4c88-9c0f-6af2dec470db-kube-api-access-sfclx" (OuterVolumeSpecName: "kube-api-access-sfclx") pod "bbafcdf6-a834-4c88-9c0f-6af2dec470db" (UID: "bbafcdf6-a834-4c88-9c0f-6af2dec470db"). InnerVolumeSpecName "kube-api-access-sfclx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.375003 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-663d-account-create-vr8wb"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.408713 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbafcdf6-a834-4c88-9c0f-6af2dec470db" (UID: "bbafcdf6-a834-4c88-9c0f-6af2dec470db"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.469780 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsr4l\" (UniqueName: \"kubernetes.io/projected/d75a804a-5ed9-4f5c-8426-85d49905f065-kube-api-access-jsr4l\") pod \"nova-api-663d-account-create-vr8wb\" (UID: \"d75a804a-5ed9-4f5c-8426-85d49905f065\") " pod="openstack/nova-api-663d-account-create-vr8wb" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.469940 4992 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.469957 4992 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.469969 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.469981 4992 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbafcdf6-a834-4c88-9c0f-6af2dec470db-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.469994 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfclx\" (UniqueName: \"kubernetes.io/projected/bbafcdf6-a834-4c88-9c0f-6af2dec470db-kube-api-access-sfclx\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.478270 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbafcdf6-a834-4c88-9c0f-6af2dec470db" (UID: "bbafcdf6-a834-4c88-9c0f-6af2dec470db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.490246 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-config-data" (OuterVolumeSpecName: "config-data") pod "bbafcdf6-a834-4c88-9c0f-6af2dec470db" (UID: "bbafcdf6-a834-4c88-9c0f-6af2dec470db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.555174 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0d44-account-create-dr6kn"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.556777 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0d44-account-create-dr6kn" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.560149 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.571104 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsr4l\" (UniqueName: \"kubernetes.io/projected/d75a804a-5ed9-4f5c-8426-85d49905f065-kube-api-access-jsr4l\") pod \"nova-api-663d-account-create-vr8wb\" (UID: \"d75a804a-5ed9-4f5c-8426-85d49905f065\") " pod="openstack/nova-api-663d-account-create-vr8wb" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.571209 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.571226 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbafcdf6-a834-4c88-9c0f-6af2dec470db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.573077 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0d44-account-create-dr6kn"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.595256 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsr4l\" (UniqueName: \"kubernetes.io/projected/d75a804a-5ed9-4f5c-8426-85d49905f065-kube-api-access-jsr4l\") pod \"nova-api-663d-account-create-vr8wb\" (UID: \"d75a804a-5ed9-4f5c-8426-85d49905f065\") " pod="openstack/nova-api-663d-account-create-vr8wb" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.673515 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9lsb\" (UniqueName: \"kubernetes.io/projected/c0298f1a-b260-4298-8c05-ffc8ac1b2d08-kube-api-access-t9lsb\") pod \"nova-cell0-0d44-account-create-dr6kn\" (UID: \"c0298f1a-b260-4298-8c05-ffc8ac1b2d08\") " pod="openstack/nova-cell0-0d44-account-create-dr6kn" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.758098 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1181-account-create-7s4m2"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.759663 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1181-account-create-7s4m2" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.762505 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.764939 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1181-account-create-7s4m2"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.775567 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9lsb\" (UniqueName: \"kubernetes.io/projected/c0298f1a-b260-4298-8c05-ffc8ac1b2d08-kube-api-access-t9lsb\") pod \"nova-cell0-0d44-account-create-dr6kn\" (UID: \"c0298f1a-b260-4298-8c05-ffc8ac1b2d08\") " pod="openstack/nova-cell0-0d44-account-create-dr6kn" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.795458 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9lsb\" (UniqueName: \"kubernetes.io/projected/c0298f1a-b260-4298-8c05-ffc8ac1b2d08-kube-api-access-t9lsb\") pod \"nova-cell0-0d44-account-create-dr6kn\" (UID: \"c0298f1a-b260-4298-8c05-ffc8ac1b2d08\") " pod="openstack/nova-cell0-0d44-account-create-dr6kn" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.817724 4992 generic.go:334] "Generic (PLEG): container finished" podID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" containerID="943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754" exitCode=0 Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.817800 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.817848 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerDied","Data":"943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754"} Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.817924 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbafcdf6-a834-4c88-9c0f-6af2dec470db","Type":"ContainerDied","Data":"4d17c4488d400c3d34c4d481e9e7f495b4e0813ee481ecd3bb8a61c5946dbef3"} Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.817943 4992 scope.go:117] "RemoveContainer" containerID="3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.829715 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-663d-account-create-vr8wb" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.849684 4992 scope.go:117] "RemoveContainer" containerID="4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.864380 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.872469 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0d44-account-create-dr6kn" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.877678 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6rll\" (UniqueName: \"kubernetes.io/projected/9ab7a5be-31da-432b-9ceb-94b52df3932b-kube-api-access-q6rll\") pod \"nova-cell1-1181-account-create-7s4m2\" (UID: \"9ab7a5be-31da-432b-9ceb-94b52df3932b\") " pod="openstack/nova-cell1-1181-account-create-7s4m2" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.878007 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.887445 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.890592 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.894476 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.894546 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.903704 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.921645 4992 scope.go:117] "RemoveContainer" containerID="0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.953675 4992 scope.go:117] "RemoveContainer" containerID="943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.978915 4992 scope.go:117] "RemoveContainer" containerID="3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.980229 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-config-data\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.980320 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.980348 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-log-httpd\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.980384 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.980433 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-scripts\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.980468 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm99d\" (UniqueName: \"kubernetes.io/projected/40e3d744-fa2c-415a-b608-126f708f9bcc-kube-api-access-jm99d\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.980574 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6rll\" (UniqueName: \"kubernetes.io/projected/9ab7a5be-31da-432b-9ceb-94b52df3932b-kube-api-access-q6rll\") pod \"nova-cell1-1181-account-create-7s4m2\" (UID: \"9ab7a5be-31da-432b-9ceb-94b52df3932b\") " pod="openstack/nova-cell1-1181-account-create-7s4m2" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.980625 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-run-httpd\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.980959 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187\": container with ID starting with 3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187 not found: ID does not exist" containerID="3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.981003 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187"} err="failed to get container status \"3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187\": rpc error: code = NotFound desc = could not find container \"3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187\": container with ID starting with 3d6d9684223838c71c77d1ffbfb1404498f2cfd8af39ce992b4f123fe1167187 not found: ID does not exist" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.981039 4992 scope.go:117] "RemoveContainer" containerID="4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.981350 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95\": container with ID starting with 4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95 not found: ID does not exist" containerID="4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.981426 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95"} err="failed to get container status \"4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95\": rpc error: code = NotFound desc = could not find container \"4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95\": container with ID starting with 4beb57d30bee5886c4c8d17078a1abdba0467493e8c746d1ae3cc48a7fddda95 not found: ID does not exist" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.981444 4992 scope.go:117] "RemoveContainer" containerID="0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.983419 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6\": container with ID starting with 0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6 not found: ID does not exist" containerID="0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.983459 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6"} err="failed to get container status \"0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6\": rpc error: code = NotFound desc = could not find container \"0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6\": container with ID starting with 0d0071c97a8775d748ffdf4ae51797fff125642dbe5a6a08b712c9db9efb45a6 not found: ID does not exist" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.983487 4992 scope.go:117] "RemoveContainer" containerID="943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754" Oct 04 04:54:47 crc kubenswrapper[4992]: E1004 04:54:47.983812 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754\": container with ID starting with 943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754 not found: ID does not exist" containerID="943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.983859 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754"} err="failed to get container status \"943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754\": rpc error: code = NotFound desc = could not find container \"943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754\": container with ID starting with 943f60dc9b7410a272f6fa02d36eba4f94872386d36776921ec40eaed02f3754 not found: ID does not exist" Oct 04 04:54:47 crc kubenswrapper[4992]: I1004 04:54:47.999192 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6rll\" (UniqueName: \"kubernetes.io/projected/9ab7a5be-31da-432b-9ceb-94b52df3932b-kube-api-access-q6rll\") pod \"nova-cell1-1181-account-create-7s4m2\" (UID: \"9ab7a5be-31da-432b-9ceb-94b52df3932b\") " pod="openstack/nova-cell1-1181-account-create-7s4m2" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.085463 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-run-httpd\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.085846 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-config-data\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.085926 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.085954 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-log-httpd\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.086121 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.086176 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-scripts\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.086212 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm99d\" (UniqueName: \"kubernetes.io/projected/40e3d744-fa2c-415a-b608-126f708f9bcc-kube-api-access-jm99d\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.086881 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-run-httpd\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.086889 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-log-httpd\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.089730 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1181-account-create-7s4m2" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.092148 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.092634 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-scripts\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.096687 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.106169 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-config-data\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.113819 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm99d\" (UniqueName: \"kubernetes.io/projected/40e3d744-fa2c-415a-b608-126f708f9bcc-kube-api-access-jm99d\") pod \"ceilometer-0\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.219209 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.332109 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbafcdf6-a834-4c88-9c0f-6af2dec470db" path="/var/lib/kubelet/pods/bbafcdf6-a834-4c88-9c0f-6af2dec470db/volumes" Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.399098 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-663d-account-create-vr8wb"] Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.478675 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0d44-account-create-dr6kn"] Oct 04 04:54:48 crc kubenswrapper[4992]: W1004 04:54:48.481111 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0298f1a_b260_4298_8c05_ffc8ac1b2d08.slice/crio-26a327532f762c74dc96b02eab1cf01a60097ae91827a849fd44539500352537 WatchSource:0}: Error finding container 26a327532f762c74dc96b02eab1cf01a60097ae91827a849fd44539500352537: Status 404 returned error can't find the container with id 26a327532f762c74dc96b02eab1cf01a60097ae91827a849fd44539500352537 Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.654084 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1181-account-create-7s4m2"] Oct 04 04:54:48 crc kubenswrapper[4992]: W1004 04:54:48.660420 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ab7a5be_31da_432b_9ceb_94b52df3932b.slice/crio-894fff02d9aecd1090b1d8f2818408fa24e9e84ef33ba29b07d445b8a353f5d3 WatchSource:0}: Error finding container 894fff02d9aecd1090b1d8f2818408fa24e9e84ef33ba29b07d445b8a353f5d3: Status 404 returned error can't find the container with id 894fff02d9aecd1090b1d8f2818408fa24e9e84ef33ba29b07d445b8a353f5d3 Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.779138 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.827391 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1181-account-create-7s4m2" event={"ID":"9ab7a5be-31da-432b-9ceb-94b52df3932b","Type":"ContainerStarted","Data":"894fff02d9aecd1090b1d8f2818408fa24e9e84ef33ba29b07d445b8a353f5d3"} Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.828689 4992 generic.go:334] "Generic (PLEG): container finished" podID="d75a804a-5ed9-4f5c-8426-85d49905f065" containerID="e83cd842770012a1ea1f2aa1079f4f2eadf00df7782980e6091c81a4da41160a" exitCode=0 Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.828732 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-663d-account-create-vr8wb" event={"ID":"d75a804a-5ed9-4f5c-8426-85d49905f065","Type":"ContainerDied","Data":"e83cd842770012a1ea1f2aa1079f4f2eadf00df7782980e6091c81a4da41160a"} Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.828785 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-663d-account-create-vr8wb" event={"ID":"d75a804a-5ed9-4f5c-8426-85d49905f065","Type":"ContainerStarted","Data":"a04619f24ebaa578ca2d9b89bb7680293fcb3b46884845da68bb9924fb699acf"} Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.833045 4992 generic.go:334] "Generic (PLEG): container finished" podID="c0298f1a-b260-4298-8c05-ffc8ac1b2d08" containerID="8cff7f33ca8498f10550ca9458b0415198e1e1b938303ee092bf4955ebc4c8b8" exitCode=0 Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.833084 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0d44-account-create-dr6kn" event={"ID":"c0298f1a-b260-4298-8c05-ffc8ac1b2d08","Type":"ContainerDied","Data":"8cff7f33ca8498f10550ca9458b0415198e1e1b938303ee092bf4955ebc4c8b8"} Oct 04 04:54:48 crc kubenswrapper[4992]: I1004 04:54:48.833112 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0d44-account-create-dr6kn" event={"ID":"c0298f1a-b260-4298-8c05-ffc8ac1b2d08","Type":"ContainerStarted","Data":"26a327532f762c74dc96b02eab1cf01a60097ae91827a849fd44539500352537"} Oct 04 04:54:48 crc kubenswrapper[4992]: W1004 04:54:48.851139 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40e3d744_fa2c_415a_b608_126f708f9bcc.slice/crio-6867682ed3bba63214e859e1edecbda42376a3a84ade0b80e42448c4586eb9e5 WatchSource:0}: Error finding container 6867682ed3bba63214e859e1edecbda42376a3a84ade0b80e42448c4586eb9e5: Status 404 returned error can't find the container with id 6867682ed3bba63214e859e1edecbda42376a3a84ade0b80e42448c4586eb9e5 Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.192572 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8699c8bbf-h2hwx" Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.257178 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ddf7f7f7b-dtwqh"] Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.257492 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-ddf7f7f7b-dtwqh" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerName="neutron-api" containerID="cri-o://fc353a7dda606f02c0bd0b3e4311fc9621bc9b49fc831576d3ec57048058ed9c" gracePeriod=30 Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.258006 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-ddf7f7f7b-dtwqh" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerName="neutron-httpd" containerID="cri-o://29c2ed4603c34fd34e9b934e9245b45af9e552ee72c0bb1b0366afc2cbac49ee" gracePeriod=30 Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.842787 4992 generic.go:334] "Generic (PLEG): container finished" podID="9ab7a5be-31da-432b-9ceb-94b52df3932b" containerID="3a3061ba0c0c8aa15a6af5f3ff733de50ef6020a7a100ca1aa4a939fb027c7af" exitCode=0 Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.842931 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1181-account-create-7s4m2" event={"ID":"9ab7a5be-31da-432b-9ceb-94b52df3932b","Type":"ContainerDied","Data":"3a3061ba0c0c8aa15a6af5f3ff733de50ef6020a7a100ca1aa4a939fb027c7af"} Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.848247 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerStarted","Data":"4ed41564ca460d4e18b4bb8ec476856767fc2f4c83441de3b989684b32d9fea8"} Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.848285 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerStarted","Data":"6867682ed3bba63214e859e1edecbda42376a3a84ade0b80e42448c4586eb9e5"} Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.850385 4992 generic.go:334] "Generic (PLEG): container finished" podID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerID="29c2ed4603c34fd34e9b934e9245b45af9e552ee72c0bb1b0366afc2cbac49ee" exitCode=0 Oct 04 04:54:49 crc kubenswrapper[4992]: I1004 04:54:49.850462 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ddf7f7f7b-dtwqh" event={"ID":"fffadcba-aa98-4d56-8b1f-86dc2435f8d0","Type":"ContainerDied","Data":"29c2ed4603c34fd34e9b934e9245b45af9e552ee72c0bb1b0366afc2cbac49ee"} Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.405709 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0d44-account-create-dr6kn" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.421551 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-663d-account-create-vr8wb" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.538771 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsr4l\" (UniqueName: \"kubernetes.io/projected/d75a804a-5ed9-4f5c-8426-85d49905f065-kube-api-access-jsr4l\") pod \"d75a804a-5ed9-4f5c-8426-85d49905f065\" (UID: \"d75a804a-5ed9-4f5c-8426-85d49905f065\") " Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.539201 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9lsb\" (UniqueName: \"kubernetes.io/projected/c0298f1a-b260-4298-8c05-ffc8ac1b2d08-kube-api-access-t9lsb\") pod \"c0298f1a-b260-4298-8c05-ffc8ac1b2d08\" (UID: \"c0298f1a-b260-4298-8c05-ffc8ac1b2d08\") " Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.544653 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d75a804a-5ed9-4f5c-8426-85d49905f065-kube-api-access-jsr4l" (OuterVolumeSpecName: "kube-api-access-jsr4l") pod "d75a804a-5ed9-4f5c-8426-85d49905f065" (UID: "d75a804a-5ed9-4f5c-8426-85d49905f065"). InnerVolumeSpecName "kube-api-access-jsr4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.547524 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0298f1a-b260-4298-8c05-ffc8ac1b2d08-kube-api-access-t9lsb" (OuterVolumeSpecName: "kube-api-access-t9lsb") pod "c0298f1a-b260-4298-8c05-ffc8ac1b2d08" (UID: "c0298f1a-b260-4298-8c05-ffc8ac1b2d08"). InnerVolumeSpecName "kube-api-access-t9lsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.641064 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsr4l\" (UniqueName: \"kubernetes.io/projected/d75a804a-5ed9-4f5c-8426-85d49905f065-kube-api-access-jsr4l\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.641106 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9lsb\" (UniqueName: \"kubernetes.io/projected/c0298f1a-b260-4298-8c05-ffc8ac1b2d08-kube-api-access-t9lsb\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.861920 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0d44-account-create-dr6kn" event={"ID":"c0298f1a-b260-4298-8c05-ffc8ac1b2d08","Type":"ContainerDied","Data":"26a327532f762c74dc96b02eab1cf01a60097ae91827a849fd44539500352537"} Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.861960 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26a327532f762c74dc96b02eab1cf01a60097ae91827a849fd44539500352537" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.862000 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0d44-account-create-dr6kn" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.864911 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerStarted","Data":"2698fd990afb8c61c9c28063203723b53d43fe3755d956834d76d61b635ade7d"} Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.867263 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-663d-account-create-vr8wb" Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.871483 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-663d-account-create-vr8wb" event={"ID":"d75a804a-5ed9-4f5c-8426-85d49905f065","Type":"ContainerDied","Data":"a04619f24ebaa578ca2d9b89bb7680293fcb3b46884845da68bb9924fb699acf"} Oct 04 04:54:50 crc kubenswrapper[4992]: I1004 04:54:50.871540 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a04619f24ebaa578ca2d9b89bb7680293fcb3b46884845da68bb9924fb699acf" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.355701 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1181-account-create-7s4m2" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.458749 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6rll\" (UniqueName: \"kubernetes.io/projected/9ab7a5be-31da-432b-9ceb-94b52df3932b-kube-api-access-q6rll\") pod \"9ab7a5be-31da-432b-9ceb-94b52df3932b\" (UID: \"9ab7a5be-31da-432b-9ceb-94b52df3932b\") " Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.471232 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab7a5be-31da-432b-9ceb-94b52df3932b-kube-api-access-q6rll" (OuterVolumeSpecName: "kube-api-access-q6rll") pod "9ab7a5be-31da-432b-9ceb-94b52df3932b" (UID: "9ab7a5be-31da-432b-9ceb-94b52df3932b"). InnerVolumeSpecName "kube-api-access-q6rll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.561547 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6rll\" (UniqueName: \"kubernetes.io/projected/9ab7a5be-31da-432b-9ceb-94b52df3932b-kube-api-access-q6rll\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.572979 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.573469 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-httpd" containerID="cri-o://1623e13998770c5dceaaa0253008e74fad9abf73c5570330dba759a257ad56da" gracePeriod=30 Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.573288 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-log" containerID="cri-o://b92e82a510849611c9aa4065a21358500e150e0b4e146d2526ecfe2f61ae942b" gracePeriod=30 Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.582907 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": EOF" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.583026 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": EOF" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.583849 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/glance-default-internal-api-0" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": EOF" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.876161 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1181-account-create-7s4m2" event={"ID":"9ab7a5be-31da-432b-9ceb-94b52df3932b","Type":"ContainerDied","Data":"894fff02d9aecd1090b1d8f2818408fa24e9e84ef33ba29b07d445b8a353f5d3"} Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.876210 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="894fff02d9aecd1090b1d8f2818408fa24e9e84ef33ba29b07d445b8a353f5d3" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.876479 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1181-account-create-7s4m2" Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.878397 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerStarted","Data":"3a94ed92c46760288607c538022ccecbabccf2632636afdac2f4df23f40f9969"} Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.880429 4992 generic.go:334] "Generic (PLEG): container finished" podID="3539e12b-5589-449d-9600-02d7aab34e88" containerID="b92e82a510849611c9aa4065a21358500e150e0b4e146d2526ecfe2f61ae942b" exitCode=143 Oct 04 04:54:51 crc kubenswrapper[4992]: I1004 04:54:51.880479 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3539e12b-5589-449d-9600-02d7aab34e88","Type":"ContainerDied","Data":"b92e82a510849611c9aa4065a21358500e150e0b4e146d2526ecfe2f61ae942b"} Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.929959 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerStarted","Data":"76a38de134c3fc765d73b5086300edd3bd36d0fe9ca6369d783ae26262ee3a17"} Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.931101 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.934419 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ctg5r"] Oct 04 04:54:52 crc kubenswrapper[4992]: E1004 04:54:52.934833 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75a804a-5ed9-4f5c-8426-85d49905f065" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.934896 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75a804a-5ed9-4f5c-8426-85d49905f065" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: E1004 04:54:52.934983 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0298f1a-b260-4298-8c05-ffc8ac1b2d08" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.935742 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0298f1a-b260-4298-8c05-ffc8ac1b2d08" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: E1004 04:54:52.935822 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab7a5be-31da-432b-9ceb-94b52df3932b" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.935873 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab7a5be-31da-432b-9ceb-94b52df3932b" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.936106 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab7a5be-31da-432b-9ceb-94b52df3932b" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.936200 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="d75a804a-5ed9-4f5c-8426-85d49905f065" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.936291 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0298f1a-b260-4298-8c05-ffc8ac1b2d08" containerName="mariadb-account-create" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.936914 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.938266 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.938585 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wx5zz" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.938684 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.943430 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ctg5r"] Oct 04 04:54:52 crc kubenswrapper[4992]: I1004 04:54:52.959257 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.408578834 podStartE2EDuration="5.959235426s" podCreationTimestamp="2025-10-04 04:54:47 +0000 UTC" firstStartedPulling="2025-10-04 04:54:48.856765955 +0000 UTC m=+1322.704441423" lastFinishedPulling="2025-10-04 04:54:52.407422547 +0000 UTC m=+1326.255098015" observedRunningTime="2025-10-04 04:54:52.95150754 +0000 UTC m=+1326.799183008" watchObservedRunningTime="2025-10-04 04:54:52.959235426 +0000 UTC m=+1326.806910894" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.020393 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.020543 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-scripts\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.020611 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-config-data\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.020676 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2wqc\" (UniqueName: \"kubernetes.io/projected/2dbd79c9-221a-4665-9046-9c8fb90cfc04-kube-api-access-z2wqc\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.122319 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-scripts\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.122425 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-config-data\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.122469 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2wqc\" (UniqueName: \"kubernetes.io/projected/2dbd79c9-221a-4665-9046-9c8fb90cfc04-kube-api-access-z2wqc\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.122503 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.130290 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-scripts\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.134017 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-config-data\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.144270 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.152405 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2wqc\" (UniqueName: \"kubernetes.io/projected/2dbd79c9-221a-4665-9046-9c8fb90cfc04-kube-api-access-z2wqc\") pod \"nova-cell0-conductor-db-sync-ctg5r\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.252079 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.714157 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ctg5r"] Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.942801 4992 generic.go:334] "Generic (PLEG): container finished" podID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerID="fc353a7dda606f02c0bd0b3e4311fc9621bc9b49fc831576d3ec57048058ed9c" exitCode=0 Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.942848 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ddf7f7f7b-dtwqh" event={"ID":"fffadcba-aa98-4d56-8b1f-86dc2435f8d0","Type":"ContainerDied","Data":"fc353a7dda606f02c0bd0b3e4311fc9621bc9b49fc831576d3ec57048058ed9c"} Oct 04 04:54:53 crc kubenswrapper[4992]: I1004 04:54:53.945722 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" event={"ID":"2dbd79c9-221a-4665-9046-9c8fb90cfc04","Type":"ContainerStarted","Data":"093f34a343c944a344b7d81fb74c916b685828a38c738c32b05982ea04021574"} Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.153985 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.246349 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-config\") pod \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.246473 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhtj2\" (UniqueName: \"kubernetes.io/projected/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-kube-api-access-dhtj2\") pod \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.246505 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-combined-ca-bundle\") pod \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.246539 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-ovndb-tls-certs\") pod \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.246559 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-httpd-config\") pod \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\" (UID: \"fffadcba-aa98-4d56-8b1f-86dc2435f8d0\") " Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.252055 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "fffadcba-aa98-4d56-8b1f-86dc2435f8d0" (UID: "fffadcba-aa98-4d56-8b1f-86dc2435f8d0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.252247 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-kube-api-access-dhtj2" (OuterVolumeSpecName: "kube-api-access-dhtj2") pod "fffadcba-aa98-4d56-8b1f-86dc2435f8d0" (UID: "fffadcba-aa98-4d56-8b1f-86dc2435f8d0"). InnerVolumeSpecName "kube-api-access-dhtj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.296558 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-config" (OuterVolumeSpecName: "config") pod "fffadcba-aa98-4d56-8b1f-86dc2435f8d0" (UID: "fffadcba-aa98-4d56-8b1f-86dc2435f8d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.306573 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fffadcba-aa98-4d56-8b1f-86dc2435f8d0" (UID: "fffadcba-aa98-4d56-8b1f-86dc2435f8d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.345647 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "fffadcba-aa98-4d56-8b1f-86dc2435f8d0" (UID: "fffadcba-aa98-4d56-8b1f-86dc2435f8d0"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.351872 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.351955 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhtj2\" (UniqueName: \"kubernetes.io/projected/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-kube-api-access-dhtj2\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.351977 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.351988 4992 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.351997 4992 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fffadcba-aa98-4d56-8b1f-86dc2435f8d0-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.846395 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.975442 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ddf7f7f7b-dtwqh" event={"ID":"fffadcba-aa98-4d56-8b1f-86dc2435f8d0","Type":"ContainerDied","Data":"2bd11cdc33b74ef68226026eac4a4594826f2ff40b6ab6ae2c497de6a236ecef"} Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.975494 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ddf7f7f7b-dtwqh" Oct 04 04:54:54 crc kubenswrapper[4992]: I1004 04:54:54.975532 4992 scope.go:117] "RemoveContainer" containerID="29c2ed4603c34fd34e9b934e9245b45af9e552ee72c0bb1b0366afc2cbac49ee" Oct 04 04:54:55 crc kubenswrapper[4992]: I1004 04:54:55.003423 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ddf7f7f7b-dtwqh"] Oct 04 04:54:55 crc kubenswrapper[4992]: I1004 04:54:55.011374 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ddf7f7f7b-dtwqh"] Oct 04 04:54:55 crc kubenswrapper[4992]: I1004 04:54:55.025585 4992 scope.go:117] "RemoveContainer" containerID="fc353a7dda606f02c0bd0b3e4311fc9621bc9b49fc831576d3ec57048058ed9c" Oct 04 04:54:55 crc kubenswrapper[4992]: I1004 04:54:55.985068 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="ceilometer-central-agent" containerID="cri-o://4ed41564ca460d4e18b4bb8ec476856767fc2f4c83441de3b989684b32d9fea8" gracePeriod=30 Oct 04 04:54:55 crc kubenswrapper[4992]: I1004 04:54:55.985097 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="sg-core" containerID="cri-o://3a94ed92c46760288607c538022ccecbabccf2632636afdac2f4df23f40f9969" gracePeriod=30 Oct 04 04:54:55 crc kubenswrapper[4992]: I1004 04:54:55.985124 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="proxy-httpd" containerID="cri-o://76a38de134c3fc765d73b5086300edd3bd36d0fe9ca6369d783ae26262ee3a17" gracePeriod=30 Oct 04 04:54:55 crc kubenswrapper[4992]: I1004 04:54:55.985127 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="ceilometer-notification-agent" containerID="cri-o://2698fd990afb8c61c9c28063203723b53d43fe3755d956834d76d61b635ade7d" gracePeriod=30 Oct 04 04:54:56 crc kubenswrapper[4992]: I1004 04:54:56.332021 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" path="/var/lib/kubelet/pods/fffadcba-aa98-4d56-8b1f-86dc2435f8d0/volumes" Oct 04 04:54:56 crc kubenswrapper[4992]: I1004 04:54:56.994964 4992 generic.go:334] "Generic (PLEG): container finished" podID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerID="76a38de134c3fc765d73b5086300edd3bd36d0fe9ca6369d783ae26262ee3a17" exitCode=0 Oct 04 04:54:56 crc kubenswrapper[4992]: I1004 04:54:56.995255 4992 generic.go:334] "Generic (PLEG): container finished" podID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerID="3a94ed92c46760288607c538022ccecbabccf2632636afdac2f4df23f40f9969" exitCode=2 Oct 04 04:54:56 crc kubenswrapper[4992]: I1004 04:54:56.995265 4992 generic.go:334] "Generic (PLEG): container finished" podID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerID="2698fd990afb8c61c9c28063203723b53d43fe3755d956834d76d61b635ade7d" exitCode=0 Oct 04 04:54:56 crc kubenswrapper[4992]: I1004 04:54:56.995033 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerDied","Data":"76a38de134c3fc765d73b5086300edd3bd36d0fe9ca6369d783ae26262ee3a17"} Oct 04 04:54:56 crc kubenswrapper[4992]: I1004 04:54:56.995304 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerDied","Data":"3a94ed92c46760288607c538022ccecbabccf2632636afdac2f4df23f40f9969"} Oct 04 04:54:56 crc kubenswrapper[4992]: I1004 04:54:56.995317 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerDied","Data":"2698fd990afb8c61c9c28063203723b53d43fe3755d956834d76d61b635ade7d"} Oct 04 04:54:58 crc kubenswrapper[4992]: I1004 04:54:58.006644 4992 generic.go:334] "Generic (PLEG): container finished" podID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerID="4ed41564ca460d4e18b4bb8ec476856767fc2f4c83441de3b989684b32d9fea8" exitCode=0 Oct 04 04:54:58 crc kubenswrapper[4992]: I1004 04:54:58.006816 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerDied","Data":"4ed41564ca460d4e18b4bb8ec476856767fc2f4c83441de3b989684b32d9fea8"} Oct 04 04:54:58 crc kubenswrapper[4992]: I1004 04:54:58.011208 4992 generic.go:334] "Generic (PLEG): container finished" podID="3539e12b-5589-449d-9600-02d7aab34e88" containerID="1623e13998770c5dceaaa0253008e74fad9abf73c5570330dba759a257ad56da" exitCode=0 Oct 04 04:54:58 crc kubenswrapper[4992]: I1004 04:54:58.011242 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3539e12b-5589-449d-9600-02d7aab34e88","Type":"ContainerDied","Data":"1623e13998770c5dceaaa0253008e74fad9abf73c5570330dba759a257ad56da"} Oct 04 04:54:59 crc kubenswrapper[4992]: I1004 04:54:59.802705 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:54:59 crc kubenswrapper[4992]: I1004 04:54:59.803182 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerName="glance-log" containerID="cri-o://06d326cfffb4e6048ef38f1a2b0139d587f975f8800c25fafe4f106304f25ab1" gracePeriod=30 Oct 04 04:54:59 crc kubenswrapper[4992]: I1004 04:54:59.804211 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerName="glance-httpd" containerID="cri-o://43c80d655db6f4d7323bea992428e7fe03178268abc9899c78ec35cf517536d6" gracePeriod=30 Oct 04 04:55:00 crc kubenswrapper[4992]: I1004 04:55:00.035216 4992 generic.go:334] "Generic (PLEG): container finished" podID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerID="06d326cfffb4e6048ef38f1a2b0139d587f975f8800c25fafe4f106304f25ab1" exitCode=143 Oct 04 04:55:00 crc kubenswrapper[4992]: I1004 04:55:00.035582 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f3378de-a1d7-4af4-8e08-de77091ed6c4","Type":"ContainerDied","Data":"06d326cfffb4e6048ef38f1a2b0139d587f975f8800c25fafe4f106304f25ab1"} Oct 04 04:55:01 crc kubenswrapper[4992]: I1004 04:55:01.868790 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:01 crc kubenswrapper[4992]: I1004 04:55:01.949808 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.019804 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm99d\" (UniqueName: \"kubernetes.io/projected/40e3d744-fa2c-415a-b608-126f708f9bcc-kube-api-access-jm99d\") pod \"40e3d744-fa2c-415a-b608-126f708f9bcc\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.020783 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-run-httpd\") pod \"40e3d744-fa2c-415a-b608-126f708f9bcc\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.020841 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-sg-core-conf-yaml\") pod \"40e3d744-fa2c-415a-b608-126f708f9bcc\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.020878 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"3539e12b-5589-449d-9600-02d7aab34e88\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.020938 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-log-httpd\") pod \"40e3d744-fa2c-415a-b608-126f708f9bcc\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021014 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-scripts\") pod \"3539e12b-5589-449d-9600-02d7aab34e88\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021070 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-logs\") pod \"3539e12b-5589-449d-9600-02d7aab34e88\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021109 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-config-data\") pod \"3539e12b-5589-449d-9600-02d7aab34e88\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021160 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-httpd-run\") pod \"3539e12b-5589-449d-9600-02d7aab34e88\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021188 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-combined-ca-bundle\") pod \"40e3d744-fa2c-415a-b608-126f708f9bcc\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021214 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-combined-ca-bundle\") pod \"3539e12b-5589-449d-9600-02d7aab34e88\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021245 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-scripts\") pod \"40e3d744-fa2c-415a-b608-126f708f9bcc\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021301 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-internal-tls-certs\") pod \"3539e12b-5589-449d-9600-02d7aab34e88\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021404 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llv8m\" (UniqueName: \"kubernetes.io/projected/3539e12b-5589-449d-9600-02d7aab34e88-kube-api-access-llv8m\") pod \"3539e12b-5589-449d-9600-02d7aab34e88\" (UID: \"3539e12b-5589-449d-9600-02d7aab34e88\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021429 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-config-data\") pod \"40e3d744-fa2c-415a-b608-126f708f9bcc\" (UID: \"40e3d744-fa2c-415a-b608-126f708f9bcc\") " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.021857 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "40e3d744-fa2c-415a-b608-126f708f9bcc" (UID: "40e3d744-fa2c-415a-b608-126f708f9bcc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.022313 4992 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.022317 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3539e12b-5589-449d-9600-02d7aab34e88" (UID: "3539e12b-5589-449d-9600-02d7aab34e88"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.022589 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-logs" (OuterVolumeSpecName: "logs") pod "3539e12b-5589-449d-9600-02d7aab34e88" (UID: "3539e12b-5589-449d-9600-02d7aab34e88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.025275 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "40e3d744-fa2c-415a-b608-126f708f9bcc" (UID: "40e3d744-fa2c-415a-b608-126f708f9bcc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.025689 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e3d744-fa2c-415a-b608-126f708f9bcc-kube-api-access-jm99d" (OuterVolumeSpecName: "kube-api-access-jm99d") pod "40e3d744-fa2c-415a-b608-126f708f9bcc" (UID: "40e3d744-fa2c-415a-b608-126f708f9bcc"). InnerVolumeSpecName "kube-api-access-jm99d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.026932 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-scripts" (OuterVolumeSpecName: "scripts") pod "3539e12b-5589-449d-9600-02d7aab34e88" (UID: "3539e12b-5589-449d-9600-02d7aab34e88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.035291 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "3539e12b-5589-449d-9600-02d7aab34e88" (UID: "3539e12b-5589-449d-9600-02d7aab34e88"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.038536 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-scripts" (OuterVolumeSpecName: "scripts") pod "40e3d744-fa2c-415a-b608-126f708f9bcc" (UID: "40e3d744-fa2c-415a-b608-126f708f9bcc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.043640 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3539e12b-5589-449d-9600-02d7aab34e88-kube-api-access-llv8m" (OuterVolumeSpecName: "kube-api-access-llv8m") pod "3539e12b-5589-449d-9600-02d7aab34e88" (UID: "3539e12b-5589-449d-9600-02d7aab34e88"). InnerVolumeSpecName "kube-api-access-llv8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.070743 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "40e3d744-fa2c-415a-b608-126f708f9bcc" (UID: "40e3d744-fa2c-415a-b608-126f708f9bcc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.078071 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3539e12b-5589-449d-9600-02d7aab34e88","Type":"ContainerDied","Data":"fecf8a58f9bff432b5b6c40dbe3ebcf743a736d9f4a7f43d09d939897099a2c8"} Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.078135 4992 scope.go:117] "RemoveContainer" containerID="1623e13998770c5dceaaa0253008e74fad9abf73c5570330dba759a257ad56da" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.078133 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.087266 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3539e12b-5589-449d-9600-02d7aab34e88" (UID: "3539e12b-5589-449d-9600-02d7aab34e88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.101747 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40e3d744-fa2c-415a-b608-126f708f9bcc","Type":"ContainerDied","Data":"6867682ed3bba63214e859e1edecbda42376a3a84ade0b80e42448c4586eb9e5"} Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.101774 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.106380 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" event={"ID":"2dbd79c9-221a-4665-9046-9c8fb90cfc04","Type":"ContainerStarted","Data":"c5797424d349d2b1dbc6ee1f637579fd4ab0bd96477674c205fea64cb5bd6d88"} Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.110475 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-config-data" (OuterVolumeSpecName: "config-data") pod "3539e12b-5589-449d-9600-02d7aab34e88" (UID: "3539e12b-5589-449d-9600-02d7aab34e88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129219 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129278 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129289 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129301 4992 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3539e12b-5589-449d-9600-02d7aab34e88-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129315 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129327 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129337 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llv8m\" (UniqueName: \"kubernetes.io/projected/3539e12b-5589-449d-9600-02d7aab34e88-kube-api-access-llv8m\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129346 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm99d\" (UniqueName: \"kubernetes.io/projected/40e3d744-fa2c-415a-b608-126f708f9bcc-kube-api-access-jm99d\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129369 4992 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129397 4992 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.129409 4992 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40e3d744-fa2c-415a-b608-126f708f9bcc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.138091 4992 scope.go:117] "RemoveContainer" containerID="b92e82a510849611c9aa4065a21358500e150e0b4e146d2526ecfe2f61ae942b" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.140841 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" podStartSLOduration=2.208169995 podStartE2EDuration="10.140817467s" podCreationTimestamp="2025-10-04 04:54:52 +0000 UTC" firstStartedPulling="2025-10-04 04:54:53.727163789 +0000 UTC m=+1327.574839257" lastFinishedPulling="2025-10-04 04:55:01.659811261 +0000 UTC m=+1335.507486729" observedRunningTime="2025-10-04 04:55:02.133512923 +0000 UTC m=+1335.981188391" watchObservedRunningTime="2025-10-04 04:55:02.140817467 +0000 UTC m=+1335.988492945" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.144784 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3539e12b-5589-449d-9600-02d7aab34e88" (UID: "3539e12b-5589-449d-9600-02d7aab34e88"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.162913 4992 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.172744 4992 scope.go:117] "RemoveContainer" containerID="76a38de134c3fc765d73b5086300edd3bd36d0fe9ca6369d783ae26262ee3a17" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.195133 4992 scope.go:117] "RemoveContainer" containerID="3a94ed92c46760288607c538022ccecbabccf2632636afdac2f4df23f40f9969" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.195427 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40e3d744-fa2c-415a-b608-126f708f9bcc" (UID: "40e3d744-fa2c-415a-b608-126f708f9bcc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.207709 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-config-data" (OuterVolumeSpecName: "config-data") pod "40e3d744-fa2c-415a-b608-126f708f9bcc" (UID: "40e3d744-fa2c-415a-b608-126f708f9bcc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.219131 4992 scope.go:117] "RemoveContainer" containerID="2698fd990afb8c61c9c28063203723b53d43fe3755d956834d76d61b635ade7d" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.231434 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.231481 4992 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3539e12b-5589-449d-9600-02d7aab34e88-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.231492 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40e3d744-fa2c-415a-b608-126f708f9bcc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.231502 4992 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.241319 4992 scope.go:117] "RemoveContainer" containerID="4ed41564ca460d4e18b4bb8ec476856767fc2f4c83441de3b989684b32d9fea8" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.406480 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.414396 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.429536 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.448255 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.457412 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:55:02 crc kubenswrapper[4992]: E1004 04:55:02.457910 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-log" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.457937 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-log" Oct 04 04:55:02 crc kubenswrapper[4992]: E1004 04:55:02.457962 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="proxy-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.457970 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="proxy-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: E1004 04:55:02.457992 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="ceilometer-notification-agent" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.457999 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="ceilometer-notification-agent" Oct 04 04:55:02 crc kubenswrapper[4992]: E1004 04:55:02.458018 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458025 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: E1004 04:55:02.458033 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="ceilometer-central-agent" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458040 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="ceilometer-central-agent" Oct 04 04:55:02 crc kubenswrapper[4992]: E1004 04:55:02.458060 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerName="neutron-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458067 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerName="neutron-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: E1004 04:55:02.458077 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerName="neutron-api" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458084 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerName="neutron-api" Oct 04 04:55:02 crc kubenswrapper[4992]: E1004 04:55:02.458098 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="sg-core" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458104 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="sg-core" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458296 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="ceilometer-notification-agent" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458310 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerName="neutron-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458323 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-log" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458336 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="fffadcba-aa98-4d56-8b1f-86dc2435f8d0" containerName="neutron-api" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458344 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="ceilometer-central-agent" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458355 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="sg-core" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458385 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3539e12b-5589-449d-9600-02d7aab34e88" containerName="glance-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.458395 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" containerName="proxy-httpd" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.459484 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.465460 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.467383 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.470242 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.474823 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.477946 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.481269 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.481779 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.482636 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538702 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538755 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538794 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538819 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-run-httpd\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538842 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-scripts\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538889 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-logs\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538904 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538929 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538977 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblvp\" (UniqueName: \"kubernetes.io/projected/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-kube-api-access-nblvp\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.538997 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.539050 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-log-httpd\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.539068 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.539085 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-config-data\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.539104 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlsvs\" (UniqueName: \"kubernetes.io/projected/c1d8f17d-925a-4102-9552-50169ed6bdc1-kube-api-access-wlsvs\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.539132 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640489 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640548 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640583 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640618 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640639 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-run-httpd\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640665 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-scripts\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640712 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-logs\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640734 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640763 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640824 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblvp\" (UniqueName: \"kubernetes.io/projected/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-kube-api-access-nblvp\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640850 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640917 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-log-httpd\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640938 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640963 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-config-data\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.640986 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlsvs\" (UniqueName: \"kubernetes.io/projected/c1d8f17d-925a-4102-9552-50169ed6bdc1-kube-api-access-wlsvs\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.641655 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-log-httpd\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.642090 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.642464 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.642790 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-run-httpd\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.645316 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-logs\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.645754 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-config-data\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.646910 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.647573 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.647948 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.648447 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.650650 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-scripts\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.651315 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.654204 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.662858 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblvp\" (UniqueName: \"kubernetes.io/projected/28ee8ac4-5271-4bbb-85db-2a83bf7cc04a-kube-api-access-nblvp\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.670648 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlsvs\" (UniqueName: \"kubernetes.io/projected/c1d8f17d-925a-4102-9552-50169ed6bdc1-kube-api-access-wlsvs\") pod \"ceilometer-0\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " pod="openstack/ceilometer-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.681680 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a\") " pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.793470 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:02 crc kubenswrapper[4992]: I1004 04:55:02.883201 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.118614 4992 generic.go:334] "Generic (PLEG): container finished" podID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerID="43c80d655db6f4d7323bea992428e7fe03178268abc9899c78ec35cf517536d6" exitCode=0 Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.118650 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f3378de-a1d7-4af4-8e08-de77091ed6c4","Type":"ContainerDied","Data":"43c80d655db6f4d7323bea992428e7fe03178268abc9899c78ec35cf517536d6"} Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.343492 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 04:55:03 crc kubenswrapper[4992]: W1004 04:55:03.387970 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1d8f17d_925a_4102_9552_50169ed6bdc1.slice/crio-2204483eb25b4201d17e237d34eee657e528a3543231d15af15d2d6c84635b93 WatchSource:0}: Error finding container 2204483eb25b4201d17e237d34eee657e528a3543231d15af15d2d6c84635b93: Status 404 returned error can't find the container with id 2204483eb25b4201d17e237d34eee657e528a3543231d15af15d2d6c84635b93 Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.407289 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.432706 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:55:03 crc kubenswrapper[4992]: E1004 04:55:03.553846 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfffadcba_aa98_4d56_8b1f_86dc2435f8d0.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563077 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-logs\") pod \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563175 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-scripts\") pod \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563227 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-combined-ca-bundle\") pod \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563341 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563402 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgt7k\" (UniqueName: \"kubernetes.io/projected/9f3378de-a1d7-4af4-8e08-de77091ed6c4-kube-api-access-dgt7k\") pod \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563499 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-config-data\") pod \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563590 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-public-tls-certs\") pod \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563646 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-httpd-run\") pod \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\" (UID: \"9f3378de-a1d7-4af4-8e08-de77091ed6c4\") " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.563762 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-logs" (OuterVolumeSpecName: "logs") pod "9f3378de-a1d7-4af4-8e08-de77091ed6c4" (UID: "9f3378de-a1d7-4af4-8e08-de77091ed6c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.564622 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.564953 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9f3378de-a1d7-4af4-8e08-de77091ed6c4" (UID: "9f3378de-a1d7-4af4-8e08-de77091ed6c4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.568952 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-scripts" (OuterVolumeSpecName: "scripts") pod "9f3378de-a1d7-4af4-8e08-de77091ed6c4" (UID: "9f3378de-a1d7-4af4-8e08-de77091ed6c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.576754 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "9f3378de-a1d7-4af4-8e08-de77091ed6c4" (UID: "9f3378de-a1d7-4af4-8e08-de77091ed6c4"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.577338 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f3378de-a1d7-4af4-8e08-de77091ed6c4-kube-api-access-dgt7k" (OuterVolumeSpecName: "kube-api-access-dgt7k") pod "9f3378de-a1d7-4af4-8e08-de77091ed6c4" (UID: "9f3378de-a1d7-4af4-8e08-de77091ed6c4"). InnerVolumeSpecName "kube-api-access-dgt7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.631988 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f3378de-a1d7-4af4-8e08-de77091ed6c4" (UID: "9f3378de-a1d7-4af4-8e08-de77091ed6c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.635448 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-config-data" (OuterVolumeSpecName: "config-data") pod "9f3378de-a1d7-4af4-8e08-de77091ed6c4" (UID: "9f3378de-a1d7-4af4-8e08-de77091ed6c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.644584 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9f3378de-a1d7-4af4-8e08-de77091ed6c4" (UID: "9f3378de-a1d7-4af4-8e08-de77091ed6c4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.666347 4992 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.666436 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgt7k\" (UniqueName: \"kubernetes.io/projected/9f3378de-a1d7-4af4-8e08-de77091ed6c4-kube-api-access-dgt7k\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.666454 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.666466 4992 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.666489 4992 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9f3378de-a1d7-4af4-8e08-de77091ed6c4-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.666499 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.666510 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f3378de-a1d7-4af4-8e08-de77091ed6c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.698554 4992 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 04:55:03 crc kubenswrapper[4992]: I1004 04:55:03.768732 4992 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.131133 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a","Type":"ContainerStarted","Data":"8daa89f6ca7cf40799ba70e0c8adb8029e5b5a0fcd651c6ecc98875834bc56c6"} Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.132123 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerStarted","Data":"2204483eb25b4201d17e237d34eee657e528a3543231d15af15d2d6c84635b93"} Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.133638 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9f3378de-a1d7-4af4-8e08-de77091ed6c4","Type":"ContainerDied","Data":"f36d6133b9dac2fcc7a245397e99600c10fc7fbba564f2183a8071653bc80b78"} Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.133683 4992 scope.go:117] "RemoveContainer" containerID="43c80d655db6f4d7323bea992428e7fe03178268abc9899c78ec35cf517536d6" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.133854 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.163828 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.170928 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.191930 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:55:04 crc kubenswrapper[4992]: E1004 04:55:04.192471 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerName="glance-httpd" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.192492 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerName="glance-httpd" Oct 04 04:55:04 crc kubenswrapper[4992]: E1004 04:55:04.192506 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerName="glance-log" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.192512 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerName="glance-log" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.192715 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerName="glance-log" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.192738 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" containerName="glance-httpd" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.194261 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.199082 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.199339 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.209634 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.256895 4992 scope.go:117] "RemoveContainer" containerID="06d326cfffb4e6048ef38f1a2b0139d587f975f8800c25fafe4f106304f25ab1" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.276111 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b879868-2ade-447f-abc0-c6656fb65ffe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.277406 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrfjt\" (UniqueName: \"kubernetes.io/projected/2b879868-2ade-447f-abc0-c6656fb65ffe-kube-api-access-vrfjt\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.277446 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.277494 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.277511 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.277577 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b879868-2ade-447f-abc0-c6656fb65ffe-logs\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.277671 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.277717 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.332745 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3539e12b-5589-449d-9600-02d7aab34e88" path="/var/lib/kubelet/pods/3539e12b-5589-449d-9600-02d7aab34e88/volumes" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.333766 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40e3d744-fa2c-415a-b608-126f708f9bcc" path="/var/lib/kubelet/pods/40e3d744-fa2c-415a-b608-126f708f9bcc/volumes" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.335705 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f3378de-a1d7-4af4-8e08-de77091ed6c4" path="/var/lib/kubelet/pods/9f3378de-a1d7-4af4-8e08-de77091ed6c4/volumes" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.379801 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrfjt\" (UniqueName: \"kubernetes.io/projected/2b879868-2ade-447f-abc0-c6656fb65ffe-kube-api-access-vrfjt\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.379864 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.379943 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.380088 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.379974 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.380649 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b879868-2ade-447f-abc0-c6656fb65ffe-logs\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.381034 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b879868-2ade-447f-abc0-c6656fb65ffe-logs\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.380814 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.381449 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.382548 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b879868-2ade-447f-abc0-c6656fb65ffe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.382838 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2b879868-2ade-447f-abc0-c6656fb65ffe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.385228 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-scripts\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.386154 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-config-data\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.386475 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.387435 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b879868-2ade-447f-abc0-c6656fb65ffe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.398161 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrfjt\" (UniqueName: \"kubernetes.io/projected/2b879868-2ade-447f-abc0-c6656fb65ffe-kube-api-access-vrfjt\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.411921 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"2b879868-2ade-447f-abc0-c6656fb65ffe\") " pod="openstack/glance-default-external-api-0" Oct 04 04:55:04 crc kubenswrapper[4992]: I1004 04:55:04.531571 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 04:55:05 crc kubenswrapper[4992]: W1004 04:55:05.134186 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b879868_2ade_447f_abc0_c6656fb65ffe.slice/crio-2b7bc14ae6feedbc38677cfbb98340008bfc827841bc6e83b92fe81d77bbc313 WatchSource:0}: Error finding container 2b7bc14ae6feedbc38677cfbb98340008bfc827841bc6e83b92fe81d77bbc313: Status 404 returned error can't find the container with id 2b7bc14ae6feedbc38677cfbb98340008bfc827841bc6e83b92fe81d77bbc313 Oct 04 04:55:05 crc kubenswrapper[4992]: I1004 04:55:05.149876 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b879868-2ade-447f-abc0-c6656fb65ffe","Type":"ContainerStarted","Data":"2b7bc14ae6feedbc38677cfbb98340008bfc827841bc6e83b92fe81d77bbc313"} Oct 04 04:55:05 crc kubenswrapper[4992]: I1004 04:55:05.153210 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 04:55:05 crc kubenswrapper[4992]: I1004 04:55:05.176607 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a","Type":"ContainerStarted","Data":"cda31309dec908b45e6df42ee4f0f96af0bcbf26f7fecbf39482f5e6bbf65808"} Oct 04 04:55:05 crc kubenswrapper[4992]: I1004 04:55:05.176655 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"28ee8ac4-5271-4bbb-85db-2a83bf7cc04a","Type":"ContainerStarted","Data":"9a6aa9833b29d381ed346f329772459a7d1c9a02db39147301da1a72fa5c6f2c"} Oct 04 04:55:05 crc kubenswrapper[4992]: I1004 04:55:05.179253 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerStarted","Data":"fc160cf105f667e87e6db1e662567a95f398cefcd19cd1f4776a605f45af60e4"} Oct 04 04:55:05 crc kubenswrapper[4992]: I1004 04:55:05.228466 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.22841471 podStartE2EDuration="3.22841471s" podCreationTimestamp="2025-10-04 04:55:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:05.212342273 +0000 UTC m=+1339.060017761" watchObservedRunningTime="2025-10-04 04:55:05.22841471 +0000 UTC m=+1339.076090188" Oct 04 04:55:07 crc kubenswrapper[4992]: I1004 04:55:07.204831 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b879868-2ade-447f-abc0-c6656fb65ffe","Type":"ContainerStarted","Data":"b1ee56124135bd9a815ed4c72f97905f57bfd9f24d8a9422c70afb48a8c3819f"} Oct 04 04:55:07 crc kubenswrapper[4992]: I1004 04:55:07.205460 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2b879868-2ade-447f-abc0-c6656fb65ffe","Type":"ContainerStarted","Data":"624c30deaee8793f08ab41c364794b8f3b7481ca56876cb0d80527cddbbf42b5"} Oct 04 04:55:07 crc kubenswrapper[4992]: I1004 04:55:07.209902 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerStarted","Data":"577c37291ac27660ec2974feb27050406829627d414a3dce008cd3968739a369"} Oct 04 04:55:07 crc kubenswrapper[4992]: I1004 04:55:07.209941 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerStarted","Data":"dc88b2dc61a04d9c6516382188481a8dcf0be778e47440a6da2bb1803f6994b6"} Oct 04 04:55:07 crc kubenswrapper[4992]: I1004 04:55:07.230249 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.230228383 podStartE2EDuration="3.230228383s" podCreationTimestamp="2025-10-04 04:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:07.225917548 +0000 UTC m=+1341.073593016" watchObservedRunningTime="2025-10-04 04:55:07.230228383 +0000 UTC m=+1341.077903851" Oct 04 04:55:09 crc kubenswrapper[4992]: I1004 04:55:09.234918 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerStarted","Data":"5914392225fc12e546252aaf3f097f9ad67ccb0db154afeed16180d4f4a71a80"} Oct 04 04:55:09 crc kubenswrapper[4992]: I1004 04:55:09.235473 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:55:09 crc kubenswrapper[4992]: I1004 04:55:09.267064 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.485916753 podStartE2EDuration="7.267043414s" podCreationTimestamp="2025-10-04 04:55:02 +0000 UTC" firstStartedPulling="2025-10-04 04:55:03.390628259 +0000 UTC m=+1337.238303727" lastFinishedPulling="2025-10-04 04:55:08.17175492 +0000 UTC m=+1342.019430388" observedRunningTime="2025-10-04 04:55:09.257790748 +0000 UTC m=+1343.105466246" watchObservedRunningTime="2025-10-04 04:55:09.267043414 +0000 UTC m=+1343.114718882" Oct 04 04:55:10 crc kubenswrapper[4992]: I1004 04:55:10.692338 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:11 crc kubenswrapper[4992]: I1004 04:55:11.254555 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="ceilometer-central-agent" containerID="cri-o://fc160cf105f667e87e6db1e662567a95f398cefcd19cd1f4776a605f45af60e4" gracePeriod=30 Oct 04 04:55:11 crc kubenswrapper[4992]: I1004 04:55:11.254582 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="proxy-httpd" containerID="cri-o://5914392225fc12e546252aaf3f097f9ad67ccb0db154afeed16180d4f4a71a80" gracePeriod=30 Oct 04 04:55:11 crc kubenswrapper[4992]: I1004 04:55:11.254607 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="sg-core" containerID="cri-o://577c37291ac27660ec2974feb27050406829627d414a3dce008cd3968739a369" gracePeriod=30 Oct 04 04:55:11 crc kubenswrapper[4992]: I1004 04:55:11.254619 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="ceilometer-notification-agent" containerID="cri-o://dc88b2dc61a04d9c6516382188481a8dcf0be778e47440a6da2bb1803f6994b6" gracePeriod=30 Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.264656 4992 generic.go:334] "Generic (PLEG): container finished" podID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerID="5914392225fc12e546252aaf3f097f9ad67ccb0db154afeed16180d4f4a71a80" exitCode=0 Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.264980 4992 generic.go:334] "Generic (PLEG): container finished" podID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerID="577c37291ac27660ec2974feb27050406829627d414a3dce008cd3968739a369" exitCode=2 Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.264995 4992 generic.go:334] "Generic (PLEG): container finished" podID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerID="dc88b2dc61a04d9c6516382188481a8dcf0be778e47440a6da2bb1803f6994b6" exitCode=0 Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.264726 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerDied","Data":"5914392225fc12e546252aaf3f097f9ad67ccb0db154afeed16180d4f4a71a80"} Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.265038 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerDied","Data":"577c37291ac27660ec2974feb27050406829627d414a3dce008cd3968739a369"} Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.265061 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerDied","Data":"dc88b2dc61a04d9c6516382188481a8dcf0be778e47440a6da2bb1803f6994b6"} Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.793781 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.793873 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.829063 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:12 crc kubenswrapper[4992]: I1004 04:55:12.843763 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.275845 4992 generic.go:334] "Generic (PLEG): container finished" podID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerID="fc160cf105f667e87e6db1e662567a95f398cefcd19cd1f4776a605f45af60e4" exitCode=0 Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.275943 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerDied","Data":"fc160cf105f667e87e6db1e662567a95f398cefcd19cd1f4776a605f45af60e4"} Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.276669 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.276705 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.409243 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.583466 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlsvs\" (UniqueName: \"kubernetes.io/projected/c1d8f17d-925a-4102-9552-50169ed6bdc1-kube-api-access-wlsvs\") pod \"c1d8f17d-925a-4102-9552-50169ed6bdc1\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.583572 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-scripts\") pod \"c1d8f17d-925a-4102-9552-50169ed6bdc1\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.583605 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-config-data\") pod \"c1d8f17d-925a-4102-9552-50169ed6bdc1\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.583643 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-sg-core-conf-yaml\") pod \"c1d8f17d-925a-4102-9552-50169ed6bdc1\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.583760 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-log-httpd\") pod \"c1d8f17d-925a-4102-9552-50169ed6bdc1\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.583794 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-combined-ca-bundle\") pod \"c1d8f17d-925a-4102-9552-50169ed6bdc1\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.583820 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-run-httpd\") pod \"c1d8f17d-925a-4102-9552-50169ed6bdc1\" (UID: \"c1d8f17d-925a-4102-9552-50169ed6bdc1\") " Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.584222 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c1d8f17d-925a-4102-9552-50169ed6bdc1" (UID: "c1d8f17d-925a-4102-9552-50169ed6bdc1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.584811 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c1d8f17d-925a-4102-9552-50169ed6bdc1" (UID: "c1d8f17d-925a-4102-9552-50169ed6bdc1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.584886 4992 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.584900 4992 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c1d8f17d-925a-4102-9552-50169ed6bdc1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.588963 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1d8f17d-925a-4102-9552-50169ed6bdc1-kube-api-access-wlsvs" (OuterVolumeSpecName: "kube-api-access-wlsvs") pod "c1d8f17d-925a-4102-9552-50169ed6bdc1" (UID: "c1d8f17d-925a-4102-9552-50169ed6bdc1"). InnerVolumeSpecName "kube-api-access-wlsvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.589107 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-scripts" (OuterVolumeSpecName: "scripts") pod "c1d8f17d-925a-4102-9552-50169ed6bdc1" (UID: "c1d8f17d-925a-4102-9552-50169ed6bdc1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.617502 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c1d8f17d-925a-4102-9552-50169ed6bdc1" (UID: "c1d8f17d-925a-4102-9552-50169ed6bdc1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.674174 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1d8f17d-925a-4102-9552-50169ed6bdc1" (UID: "c1d8f17d-925a-4102-9552-50169ed6bdc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.686760 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.686790 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlsvs\" (UniqueName: \"kubernetes.io/projected/c1d8f17d-925a-4102-9552-50169ed6bdc1-kube-api-access-wlsvs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.686800 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.686809 4992 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.720458 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-config-data" (OuterVolumeSpecName: "config-data") pod "c1d8f17d-925a-4102-9552-50169ed6bdc1" (UID: "c1d8f17d-925a-4102-9552-50169ed6bdc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:13 crc kubenswrapper[4992]: I1004 04:55:13.789014 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f17d-925a-4102-9552-50169ed6bdc1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:13 crc kubenswrapper[4992]: E1004 04:55:13.887681 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfffadcba_aa98_4d56_8b1f_86dc2435f8d0.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.287066 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c1d8f17d-925a-4102-9552-50169ed6bdc1","Type":"ContainerDied","Data":"2204483eb25b4201d17e237d34eee657e528a3543231d15af15d2d6c84635b93"} Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.287145 4992 scope.go:117] "RemoveContainer" containerID="5914392225fc12e546252aaf3f097f9ad67ccb0db154afeed16180d4f4a71a80" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.287150 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.310814 4992 scope.go:117] "RemoveContainer" containerID="577c37291ac27660ec2974feb27050406829627d414a3dce008cd3968739a369" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.333823 4992 scope.go:117] "RemoveContainer" containerID="dc88b2dc61a04d9c6516382188481a8dcf0be778e47440a6da2bb1803f6994b6" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.339132 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.350733 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.363489 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:14 crc kubenswrapper[4992]: E1004 04:55:14.364278 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="ceilometer-notification-agent" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.364402 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="ceilometer-notification-agent" Oct 04 04:55:14 crc kubenswrapper[4992]: E1004 04:55:14.364502 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="proxy-httpd" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.364579 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="proxy-httpd" Oct 04 04:55:14 crc kubenswrapper[4992]: E1004 04:55:14.364667 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="sg-core" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.364738 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="sg-core" Oct 04 04:55:14 crc kubenswrapper[4992]: E1004 04:55:14.364840 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="ceilometer-central-agent" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.364940 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="ceilometer-central-agent" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.365283 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="ceilometer-notification-agent" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.365416 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="ceilometer-central-agent" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.365515 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="proxy-httpd" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.365602 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" containerName="sg-core" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.365999 4992 scope.go:117] "RemoveContainer" containerID="fc160cf105f667e87e6db1e662567a95f398cefcd19cd1f4776a605f45af60e4" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.368346 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.371441 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.371604 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.372594 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.501734 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.501823 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-config-data\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.501853 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-log-httpd\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.501894 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.501912 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-run-httpd\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.501938 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-scripts\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.502006 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmhc2\" (UniqueName: \"kubernetes.io/projected/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-kube-api-access-kmhc2\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.532707 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.532981 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.567785 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.581472 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.603299 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.603446 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-config-data\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.603484 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-log-httpd\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.603520 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.603537 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-run-httpd\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.603563 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-scripts\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.603617 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmhc2\" (UniqueName: \"kubernetes.io/projected/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-kube-api-access-kmhc2\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.604312 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-run-httpd\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.605257 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-log-httpd\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.607713 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.608057 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-config-data\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.608959 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.611601 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-scripts\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.622625 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmhc2\" (UniqueName: \"kubernetes.io/projected/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-kube-api-access-kmhc2\") pod \"ceilometer-0\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " pod="openstack/ceilometer-0" Oct 04 04:55:14 crc kubenswrapper[4992]: I1004 04:55:14.700873 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:15 crc kubenswrapper[4992]: I1004 04:55:15.122469 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:15 crc kubenswrapper[4992]: W1004 04:55:15.124801 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeec65c5e_a9d7_42ef_9d2a_0787ab1d7320.slice/crio-ae457492d1efcac55e6b2e86e8943b74adb9e5a8da05a53581a00b130ba57667 WatchSource:0}: Error finding container ae457492d1efcac55e6b2e86e8943b74adb9e5a8da05a53581a00b130ba57667: Status 404 returned error can't find the container with id ae457492d1efcac55e6b2e86e8943b74adb9e5a8da05a53581a00b130ba57667 Oct 04 04:55:15 crc kubenswrapper[4992]: I1004 04:55:15.304052 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerStarted","Data":"ae457492d1efcac55e6b2e86e8943b74adb9e5a8da05a53581a00b130ba57667"} Oct 04 04:55:15 crc kubenswrapper[4992]: I1004 04:55:15.304125 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:55:15 crc kubenswrapper[4992]: I1004 04:55:15.304406 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 04:55:15 crc kubenswrapper[4992]: I1004 04:55:15.373103 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:15 crc kubenswrapper[4992]: I1004 04:55:15.373751 4992 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 04:55:15 crc kubenswrapper[4992]: I1004 04:55:15.374088 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 04:55:16 crc kubenswrapper[4992]: I1004 04:55:16.313097 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerStarted","Data":"51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318"} Oct 04 04:55:16 crc kubenswrapper[4992]: I1004 04:55:16.333502 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1d8f17d-925a-4102-9552-50169ed6bdc1" path="/var/lib/kubelet/pods/c1d8f17d-925a-4102-9552-50169ed6bdc1/volumes" Oct 04 04:55:17 crc kubenswrapper[4992]: I1004 04:55:17.259393 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:55:17 crc kubenswrapper[4992]: I1004 04:55:17.261269 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 04:55:17 crc kubenswrapper[4992]: I1004 04:55:17.327793 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerStarted","Data":"ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8"} Oct 04 04:55:19 crc kubenswrapper[4992]: I1004 04:55:19.354679 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerStarted","Data":"406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c"} Oct 04 04:55:21 crc kubenswrapper[4992]: I1004 04:55:21.391233 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerStarted","Data":"f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318"} Oct 04 04:55:21 crc kubenswrapper[4992]: I1004 04:55:21.391587 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:55:21 crc kubenswrapper[4992]: I1004 04:55:21.395724 4992 generic.go:334] "Generic (PLEG): container finished" podID="2dbd79c9-221a-4665-9046-9c8fb90cfc04" containerID="c5797424d349d2b1dbc6ee1f637579fd4ab0bd96477674c205fea64cb5bd6d88" exitCode=0 Oct 04 04:55:21 crc kubenswrapper[4992]: I1004 04:55:21.395792 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" event={"ID":"2dbd79c9-221a-4665-9046-9c8fb90cfc04","Type":"ContainerDied","Data":"c5797424d349d2b1dbc6ee1f637579fd4ab0bd96477674c205fea64cb5bd6d88"} Oct 04 04:55:21 crc kubenswrapper[4992]: I1004 04:55:21.413791 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.912180524 podStartE2EDuration="7.413772625s" podCreationTimestamp="2025-10-04 04:55:14 +0000 UTC" firstStartedPulling="2025-10-04 04:55:15.12730589 +0000 UTC m=+1348.974981358" lastFinishedPulling="2025-10-04 04:55:20.628897991 +0000 UTC m=+1354.476573459" observedRunningTime="2025-10-04 04:55:21.413002194 +0000 UTC m=+1355.260677662" watchObservedRunningTime="2025-10-04 04:55:21.413772625 +0000 UTC m=+1355.261448083" Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.756092 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.870262 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-scripts\") pod \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.870389 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-config-data\") pod \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.870523 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2wqc\" (UniqueName: \"kubernetes.io/projected/2dbd79c9-221a-4665-9046-9c8fb90cfc04-kube-api-access-z2wqc\") pod \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.870570 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-combined-ca-bundle\") pod \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\" (UID: \"2dbd79c9-221a-4665-9046-9c8fb90cfc04\") " Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.875860 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-scripts" (OuterVolumeSpecName: "scripts") pod "2dbd79c9-221a-4665-9046-9c8fb90cfc04" (UID: "2dbd79c9-221a-4665-9046-9c8fb90cfc04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.875885 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dbd79c9-221a-4665-9046-9c8fb90cfc04-kube-api-access-z2wqc" (OuterVolumeSpecName: "kube-api-access-z2wqc") pod "2dbd79c9-221a-4665-9046-9c8fb90cfc04" (UID: "2dbd79c9-221a-4665-9046-9c8fb90cfc04"). InnerVolumeSpecName "kube-api-access-z2wqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.902548 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-config-data" (OuterVolumeSpecName: "config-data") pod "2dbd79c9-221a-4665-9046-9c8fb90cfc04" (UID: "2dbd79c9-221a-4665-9046-9c8fb90cfc04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:22 crc kubenswrapper[4992]: I1004 04:55:22.904153 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2dbd79c9-221a-4665-9046-9c8fb90cfc04" (UID: "2dbd79c9-221a-4665-9046-9c8fb90cfc04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.001395 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.001431 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.001441 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2wqc\" (UniqueName: \"kubernetes.io/projected/2dbd79c9-221a-4665-9046-9c8fb90cfc04-kube-api-access-z2wqc\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.001454 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dbd79c9-221a-4665-9046-9c8fb90cfc04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.417422 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" event={"ID":"2dbd79c9-221a-4665-9046-9c8fb90cfc04","Type":"ContainerDied","Data":"093f34a343c944a344b7d81fb74c916b685828a38c738c32b05982ea04021574"} Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.417466 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="093f34a343c944a344b7d81fb74c916b685828a38c738c32b05982ea04021574" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.417529 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ctg5r" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.535101 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:55:23 crc kubenswrapper[4992]: E1004 04:55:23.535476 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbd79c9-221a-4665-9046-9c8fb90cfc04" containerName="nova-cell0-conductor-db-sync" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.535523 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbd79c9-221a-4665-9046-9c8fb90cfc04" containerName="nova-cell0-conductor-db-sync" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.535699 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dbd79c9-221a-4665-9046-9c8fb90cfc04" containerName="nova-cell0-conductor-db-sync" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.536275 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.540820 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wx5zz" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.540865 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.545402 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.610661 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.610740 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.610863 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6fpg\" (UniqueName: \"kubernetes.io/projected/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-kube-api-access-j6fpg\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.712626 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.712688 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.712736 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6fpg\" (UniqueName: \"kubernetes.io/projected/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-kube-api-access-j6fpg\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.717587 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.721928 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.730989 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6fpg\" (UniqueName: \"kubernetes.io/projected/f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee-kube-api-access-j6fpg\") pod \"nova-cell0-conductor-0\" (UID: \"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee\") " pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:23 crc kubenswrapper[4992]: I1004 04:55:23.869483 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:24 crc kubenswrapper[4992]: E1004 04:55:24.158201 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfffadcba_aa98_4d56_8b1f_86dc2435f8d0.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:55:24 crc kubenswrapper[4992]: I1004 04:55:24.294758 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 04:55:24 crc kubenswrapper[4992]: W1004 04:55:24.297056 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf86bdf47_6f6d_4e6d_bcb4_c61cbd7744ee.slice/crio-a7a32019739c01e32ea56ae074cd2019f349e85d1a022bdbcda6b7f6fde39f45 WatchSource:0}: Error finding container a7a32019739c01e32ea56ae074cd2019f349e85d1a022bdbcda6b7f6fde39f45: Status 404 returned error can't find the container with id a7a32019739c01e32ea56ae074cd2019f349e85d1a022bdbcda6b7f6fde39f45 Oct 04 04:55:24 crc kubenswrapper[4992]: I1004 04:55:24.428980 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee","Type":"ContainerStarted","Data":"a7a32019739c01e32ea56ae074cd2019f349e85d1a022bdbcda6b7f6fde39f45"} Oct 04 04:55:25 crc kubenswrapper[4992]: I1004 04:55:25.448282 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee","Type":"ContainerStarted","Data":"ff080de8f04caa0bf729787466d911bb4ce3c906a8c43f6a2a4948400fc6c32a"} Oct 04 04:55:25 crc kubenswrapper[4992]: I1004 04:55:25.466136 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.466118923 podStartE2EDuration="2.466118923s" podCreationTimestamp="2025-10-04 04:55:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:25.46298647 +0000 UTC m=+1359.310661938" watchObservedRunningTime="2025-10-04 04:55:25.466118923 +0000 UTC m=+1359.313794391" Oct 04 04:55:26 crc kubenswrapper[4992]: I1004 04:55:26.456671 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:33 crc kubenswrapper[4992]: I1004 04:55:33.897894 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.379903 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-b5b2w"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.382054 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.385613 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.388119 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.392462 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-b5b2w"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.496984 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.502241 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.506825 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.515288 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcgzn\" (UniqueName: \"kubernetes.io/projected/30b06dc6-6c8d-407b-bd1f-96f05c45660b-kube-api-access-wcgzn\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.515353 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-config-data\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.518087 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-scripts\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.518155 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: E1004 04:55:34.536733 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfffadcba_aa98_4d56_8b1f_86dc2435f8d0.slice\": RecentStats: unable to find data in memory cache]" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.557430 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.572976 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.574181 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.590716 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.599433 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.620060 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.620172 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb4x2\" (UniqueName: \"kubernetes.io/projected/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-kube-api-access-qb4x2\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.620257 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.620287 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-config-data\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.620332 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcgzn\" (UniqueName: \"kubernetes.io/projected/30b06dc6-6c8d-407b-bd1f-96f05c45660b-kube-api-access-wcgzn\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.620397 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-config-data\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.620467 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-scripts\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.620506 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-logs\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.652238 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.660171 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-config-data\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.670122 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcgzn\" (UniqueName: \"kubernetes.io/projected/30b06dc6-6c8d-407b-bd1f-96f05c45660b-kube-api-access-wcgzn\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.681017 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.682430 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.697018 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-scripts\") pod \"nova-cell0-cell-mapping-b5b2w\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.701125 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.713655 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.722965 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-logs\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.723548 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.723602 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb4x2\" (UniqueName: \"kubernetes.io/projected/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-kube-api-access-qb4x2\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.723626 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.723654 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjw2c\" (UniqueName: \"kubernetes.io/projected/9497ef49-7631-453f-8c40-2fe991a823fa-kube-api-access-qjw2c\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.723712 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.723737 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-config-data\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.725105 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-logs\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.727449 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.730034 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-config-data\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.732104 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.744517 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb4x2\" (UniqueName: \"kubernetes.io/projected/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-kube-api-access-qb4x2\") pod \"nova-api-0\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.801802 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.803703 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.815661 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.827534 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.827617 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.827643 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhdvn\" (UniqueName: \"kubernetes.io/projected/1fb7d1b9-f64a-4e58-92d7-f455273a7963-kube-api-access-dhdvn\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.827681 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.827699 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjw2c\" (UniqueName: \"kubernetes.io/projected/9497ef49-7631-453f-8c40-2fe991a823fa-kube-api-access-qjw2c\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.827723 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-config-data\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.835094 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.835513 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.846777 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.852342 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.874002 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjw2c\" (UniqueName: \"kubernetes.io/projected/9497ef49-7631-453f-8c40-2fe991a823fa-kube-api-access-qjw2c\") pod \"nova-cell1-novncproxy-0\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.906471 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-8xvd8"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.908664 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.920467 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.929605 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-8xvd8"] Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.933460 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7946\" (UniqueName: \"kubernetes.io/projected/68a4ac77-95df-4b11-958c-08f07fb6e801-kube-api-access-p7946\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.933565 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68a4ac77-95df-4b11-958c-08f07fb6e801-logs\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.933602 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhdvn\" (UniqueName: \"kubernetes.io/projected/1fb7d1b9-f64a-4e58-92d7-f455273a7963-kube-api-access-dhdvn\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.933676 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-config-data\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.933734 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.933816 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-config-data\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.933861 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.946140 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.948512 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-config-data\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:34 crc kubenswrapper[4992]: I1004 04:55:34.976408 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhdvn\" (UniqueName: \"kubernetes.io/projected/1fb7d1b9-f64a-4e58-92d7-f455273a7963-kube-api-access-dhdvn\") pod \"nova-scheduler-0\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035089 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035159 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035204 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035231 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-config\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035256 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035295 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-config-data\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035317 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035394 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc5lt\" (UniqueName: \"kubernetes.io/projected/7e30d629-2fa2-4a60-be4d-cc73893f5454-kube-api-access-cc5lt\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035436 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7946\" (UniqueName: \"kubernetes.io/projected/68a4ac77-95df-4b11-958c-08f07fb6e801-kube-api-access-p7946\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.035476 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68a4ac77-95df-4b11-958c-08f07fb6e801-logs\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.056436 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68a4ac77-95df-4b11-958c-08f07fb6e801-logs\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.065185 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.077320 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7946\" (UniqueName: \"kubernetes.io/projected/68a4ac77-95df-4b11-958c-08f07fb6e801-kube-api-access-p7946\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.085593 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-config-data\") pod \"nova-metadata-0\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.162198 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.162282 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.162311 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-config\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.162332 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.162393 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.162445 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc5lt\" (UniqueName: \"kubernetes.io/projected/7e30d629-2fa2-4a60-be4d-cc73893f5454-kube-api-access-cc5lt\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.163519 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-config\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.163729 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.164313 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.168042 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.168247 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.186828 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc5lt\" (UniqueName: \"kubernetes.io/projected/7e30d629-2fa2-4a60-be4d-cc73893f5454-kube-api-access-cc5lt\") pod \"dnsmasq-dns-845d6d6f59-8xvd8\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.236798 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.245205 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.298838 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.494666 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-b5b2w"] Oct 04 04:55:35 crc kubenswrapper[4992]: W1004 04:55:35.564821 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b06dc6_6c8d_407b_bd1f_96f05c45660b.slice/crio-d7f7d1c7c3a4c5081e29b1fd1811a408075748deaae1a5fdc47cf661b647bb1b WatchSource:0}: Error finding container d7f7d1c7c3a4c5081e29b1fd1811a408075748deaae1a5fdc47cf661b647bb1b: Status 404 returned error can't find the container with id d7f7d1c7c3a4c5081e29b1fd1811a408075748deaae1a5fdc47cf661b647bb1b Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.663467 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.710230 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.729265 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ts9f9"] Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.735517 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.738383 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.739149 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.743152 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ts9f9"] Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.886787 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-scripts\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.886853 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6pgt\" (UniqueName: \"kubernetes.io/projected/2cc135c0-8c87-4802-836a-8c59def0eb0d-kube-api-access-j6pgt\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.886992 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.887075 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-config-data\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.989481 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.989610 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-config-data\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.991086 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-scripts\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.991127 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6pgt\" (UniqueName: \"kubernetes.io/projected/2cc135c0-8c87-4802-836a-8c59def0eb0d-kube-api-access-j6pgt\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.996213 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-scripts\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.996464 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:35 crc kubenswrapper[4992]: I1004 04:55:35.996489 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-config-data\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.023934 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6pgt\" (UniqueName: \"kubernetes.io/projected/2cc135c0-8c87-4802-836a-8c59def0eb0d-kube-api-access-j6pgt\") pod \"nova-cell1-conductor-db-sync-ts9f9\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.069089 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.090190 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:36 crc kubenswrapper[4992]: W1004 04:55:36.101465 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fb7d1b9_f64a_4e58_92d7_f455273a7963.slice/crio-d2af1ecfce7356592fc00a246b56df8e1c3341711f0546f33db368c25a836480 WatchSource:0}: Error finding container d2af1ecfce7356592fc00a246b56df8e1c3341711f0546f33db368c25a836480: Status 404 returned error can't find the container with id d2af1ecfce7356592fc00a246b56df8e1c3341711f0546f33db368c25a836480 Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.107860 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:36 crc kubenswrapper[4992]: W1004 04:55:36.129191 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68a4ac77_95df_4b11_958c_08f07fb6e801.slice/crio-0c1dd37feb63b1a4d1ea39fd037b39ab6d62385a82820c82602f16566326fa01 WatchSource:0}: Error finding container 0c1dd37feb63b1a4d1ea39fd037b39ab6d62385a82820c82602f16566326fa01: Status 404 returned error can't find the container with id 0c1dd37feb63b1a4d1ea39fd037b39ab6d62385a82820c82602f16566326fa01 Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.256687 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-8xvd8"] Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.565397 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-b5b2w" event={"ID":"30b06dc6-6c8d-407b-bd1f-96f05c45660b","Type":"ContainerStarted","Data":"ebce72ef6dca56815743ca8f5b44109b59750eef660a14e596c39c5216e461ae"} Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.565508 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-b5b2w" event={"ID":"30b06dc6-6c8d-407b-bd1f-96f05c45660b","Type":"ContainerStarted","Data":"d7f7d1c7c3a4c5081e29b1fd1811a408075748deaae1a5fdc47cf661b647bb1b"} Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.567016 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" event={"ID":"7e30d629-2fa2-4a60-be4d-cc73893f5454","Type":"ContainerStarted","Data":"219c28379870f40df08115184b886814aaa17c59833571655b3abfb1fe7f0898"} Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.568005 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1fb7d1b9-f64a-4e58-92d7-f455273a7963","Type":"ContainerStarted","Data":"d2af1ecfce7356592fc00a246b56df8e1c3341711f0546f33db368c25a836480"} Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.569477 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9","Type":"ContainerStarted","Data":"fa04cee4feeb7543e483b79022a285ff78d0accf87ae1a7d281ef93d6009aaea"} Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.570479 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9497ef49-7631-453f-8c40-2fe991a823fa","Type":"ContainerStarted","Data":"924d5c8a46b9f6d1ad16d196c55255e3f35259edb2f2db8c282a5010597e0ae4"} Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.571734 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68a4ac77-95df-4b11-958c-08f07fb6e801","Type":"ContainerStarted","Data":"0c1dd37feb63b1a4d1ea39fd037b39ab6d62385a82820c82602f16566326fa01"} Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.586776 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-b5b2w" podStartSLOduration=2.586755318 podStartE2EDuration="2.586755318s" podCreationTimestamp="2025-10-04 04:55:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:36.579234118 +0000 UTC m=+1370.426909596" watchObservedRunningTime="2025-10-04 04:55:36.586755318 +0000 UTC m=+1370.434430786" Oct 04 04:55:36 crc kubenswrapper[4992]: I1004 04:55:36.684129 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ts9f9"] Oct 04 04:55:36 crc kubenswrapper[4992]: W1004 04:55:36.827194 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cc135c0_8c87_4802_836a_8c59def0eb0d.slice/crio-c0474efb7e7260850b9a088445c8e58a89f5087e9ff363ce534de2bd603ecf4a WatchSource:0}: Error finding container c0474efb7e7260850b9a088445c8e58a89f5087e9ff363ce534de2bd603ecf4a: Status 404 returned error can't find the container with id c0474efb7e7260850b9a088445c8e58a89f5087e9ff363ce534de2bd603ecf4a Oct 04 04:55:37 crc kubenswrapper[4992]: I1004 04:55:37.587218 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" event={"ID":"2cc135c0-8c87-4802-836a-8c59def0eb0d","Type":"ContainerStarted","Data":"1ceb7c240fb308cad875d20402abe1d34f87bdc00eb054ac85c8762e2ae05835"} Oct 04 04:55:37 crc kubenswrapper[4992]: I1004 04:55:37.587701 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" event={"ID":"2cc135c0-8c87-4802-836a-8c59def0eb0d","Type":"ContainerStarted","Data":"c0474efb7e7260850b9a088445c8e58a89f5087e9ff363ce534de2bd603ecf4a"} Oct 04 04:55:37 crc kubenswrapper[4992]: I1004 04:55:37.595207 4992 generic.go:334] "Generic (PLEG): container finished" podID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerID="fbc115b24de9a1455b159445670a382c881f3da18b8179226a8111efe1ea33c7" exitCode=0 Oct 04 04:55:37 crc kubenswrapper[4992]: I1004 04:55:37.595686 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" event={"ID":"7e30d629-2fa2-4a60-be4d-cc73893f5454","Type":"ContainerDied","Data":"fbc115b24de9a1455b159445670a382c881f3da18b8179226a8111efe1ea33c7"} Oct 04 04:55:37 crc kubenswrapper[4992]: I1004 04:55:37.608638 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" podStartSLOduration=2.60861111 podStartE2EDuration="2.60861111s" podCreationTimestamp="2025-10-04 04:55:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:37.601884311 +0000 UTC m=+1371.449559779" watchObservedRunningTime="2025-10-04 04:55:37.60861111 +0000 UTC m=+1371.456286578" Oct 04 04:55:38 crc kubenswrapper[4992]: I1004 04:55:38.471417 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:38 crc kubenswrapper[4992]: I1004 04:55:38.501433 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.631842 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68a4ac77-95df-4b11-958c-08f07fb6e801","Type":"ContainerStarted","Data":"6d705b5502e830807e782be561dd2cc95db110bc7563221f7f4e61bfd0acf329"} Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.634172 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" event={"ID":"7e30d629-2fa2-4a60-be4d-cc73893f5454","Type":"ContainerStarted","Data":"472db47216833d39a4fb8cca0b5a138ac1238adcf1e32c55b306b33fb9eaf815"} Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.634299 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.636049 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1fb7d1b9-f64a-4e58-92d7-f455273a7963","Type":"ContainerStarted","Data":"20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9"} Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.637874 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9","Type":"ContainerStarted","Data":"e409510fc0911b8e2ead155ff70612039aac686fe8c37890cefc23a417435f61"} Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.639780 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9497ef49-7631-453f-8c40-2fe991a823fa","Type":"ContainerStarted","Data":"2a2358f1c9f9d51ebe58d267edcafd4b0425263be03c3045909384bceca080a5"} Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.639898 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="9497ef49-7631-453f-8c40-2fe991a823fa" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2a2358f1c9f9d51ebe58d267edcafd4b0425263be03c3045909384bceca080a5" gracePeriod=30 Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.658776 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" podStartSLOduration=6.658754568 podStartE2EDuration="6.658754568s" podCreationTimestamp="2025-10-04 04:55:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:40.655476111 +0000 UTC m=+1374.503151579" watchObservedRunningTime="2025-10-04 04:55:40.658754568 +0000 UTC m=+1374.506430036" Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.673722 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.469173011 podStartE2EDuration="6.673704125s" podCreationTimestamp="2025-10-04 04:55:34 +0000 UTC" firstStartedPulling="2025-10-04 04:55:35.675577007 +0000 UTC m=+1369.523252465" lastFinishedPulling="2025-10-04 04:55:39.880108101 +0000 UTC m=+1373.727783579" observedRunningTime="2025-10-04 04:55:40.669830972 +0000 UTC m=+1374.517506460" watchObservedRunningTime="2025-10-04 04:55:40.673704125 +0000 UTC m=+1374.521379593" Oct 04 04:55:40 crc kubenswrapper[4992]: I1004 04:55:40.689310 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.931635435 podStartE2EDuration="6.68929233s" podCreationTimestamp="2025-10-04 04:55:34 +0000 UTC" firstStartedPulling="2025-10-04 04:55:36.117857114 +0000 UTC m=+1369.965532582" lastFinishedPulling="2025-10-04 04:55:39.875514009 +0000 UTC m=+1373.723189477" observedRunningTime="2025-10-04 04:55:40.686442664 +0000 UTC m=+1374.534118132" watchObservedRunningTime="2025-10-04 04:55:40.68929233 +0000 UTC m=+1374.536967798" Oct 04 04:55:41 crc kubenswrapper[4992]: I1004 04:55:41.670918 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9","Type":"ContainerStarted","Data":"a860dcaa514cd33c4368910e7b744e5e3a99aa0e7933bdafded92178606188ad"} Oct 04 04:55:41 crc kubenswrapper[4992]: I1004 04:55:41.673993 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68a4ac77-95df-4b11-958c-08f07fb6e801","Type":"ContainerStarted","Data":"cc195017fa8e25b4716744641b10076a70498ff9ea29eae2b951f1edc63a791b"} Oct 04 04:55:41 crc kubenswrapper[4992]: I1004 04:55:41.674415 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerName="nova-metadata-metadata" containerID="cri-o://cc195017fa8e25b4716744641b10076a70498ff9ea29eae2b951f1edc63a791b" gracePeriod=30 Oct 04 04:55:41 crc kubenswrapper[4992]: I1004 04:55:41.674412 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerName="nova-metadata-log" containerID="cri-o://6d705b5502e830807e782be561dd2cc95db110bc7563221f7f4e61bfd0acf329" gracePeriod=30 Oct 04 04:55:41 crc kubenswrapper[4992]: I1004 04:55:41.695911 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.455830099 podStartE2EDuration="7.695891717s" podCreationTimestamp="2025-10-04 04:55:34 +0000 UTC" firstStartedPulling="2025-10-04 04:55:35.636440437 +0000 UTC m=+1369.484115905" lastFinishedPulling="2025-10-04 04:55:39.876502045 +0000 UTC m=+1373.724177523" observedRunningTime="2025-10-04 04:55:41.694846019 +0000 UTC m=+1375.542521487" watchObservedRunningTime="2025-10-04 04:55:41.695891717 +0000 UTC m=+1375.543567185" Oct 04 04:55:41 crc kubenswrapper[4992]: I1004 04:55:41.723898 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.986849355 podStartE2EDuration="7.723760148s" podCreationTimestamp="2025-10-04 04:55:34 +0000 UTC" firstStartedPulling="2025-10-04 04:55:36.140836535 +0000 UTC m=+1369.988512043" lastFinishedPulling="2025-10-04 04:55:39.877747358 +0000 UTC m=+1373.725422836" observedRunningTime="2025-10-04 04:55:41.714343077 +0000 UTC m=+1375.562018545" watchObservedRunningTime="2025-10-04 04:55:41.723760148 +0000 UTC m=+1375.571435616" Oct 04 04:55:42 crc kubenswrapper[4992]: I1004 04:55:42.683695 4992 generic.go:334] "Generic (PLEG): container finished" podID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerID="cc195017fa8e25b4716744641b10076a70498ff9ea29eae2b951f1edc63a791b" exitCode=0 Oct 04 04:55:42 crc kubenswrapper[4992]: I1004 04:55:42.683738 4992 generic.go:334] "Generic (PLEG): container finished" podID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerID="6d705b5502e830807e782be561dd2cc95db110bc7563221f7f4e61bfd0acf329" exitCode=143 Oct 04 04:55:42 crc kubenswrapper[4992]: I1004 04:55:42.683797 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68a4ac77-95df-4b11-958c-08f07fb6e801","Type":"ContainerDied","Data":"cc195017fa8e25b4716744641b10076a70498ff9ea29eae2b951f1edc63a791b"} Oct 04 04:55:42 crc kubenswrapper[4992]: I1004 04:55:42.683859 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68a4ac77-95df-4b11-958c-08f07fb6e801","Type":"ContainerDied","Data":"6d705b5502e830807e782be561dd2cc95db110bc7563221f7f4e61bfd0acf329"} Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.500922 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.660170 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-combined-ca-bundle\") pod \"68a4ac77-95df-4b11-958c-08f07fb6e801\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.660394 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7946\" (UniqueName: \"kubernetes.io/projected/68a4ac77-95df-4b11-958c-08f07fb6e801-kube-api-access-p7946\") pod \"68a4ac77-95df-4b11-958c-08f07fb6e801\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.660925 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68a4ac77-95df-4b11-958c-08f07fb6e801-logs\") pod \"68a4ac77-95df-4b11-958c-08f07fb6e801\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.661128 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-config-data\") pod \"68a4ac77-95df-4b11-958c-08f07fb6e801\" (UID: \"68a4ac77-95df-4b11-958c-08f07fb6e801\") " Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.661259 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68a4ac77-95df-4b11-958c-08f07fb6e801-logs" (OuterVolumeSpecName: "logs") pod "68a4ac77-95df-4b11-958c-08f07fb6e801" (UID: "68a4ac77-95df-4b11-958c-08f07fb6e801"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.662302 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68a4ac77-95df-4b11-958c-08f07fb6e801-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.667295 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a4ac77-95df-4b11-958c-08f07fb6e801-kube-api-access-p7946" (OuterVolumeSpecName: "kube-api-access-p7946") pod "68a4ac77-95df-4b11-958c-08f07fb6e801" (UID: "68a4ac77-95df-4b11-958c-08f07fb6e801"). InnerVolumeSpecName "kube-api-access-p7946". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.695323 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-config-data" (OuterVolumeSpecName: "config-data") pod "68a4ac77-95df-4b11-958c-08f07fb6e801" (UID: "68a4ac77-95df-4b11-958c-08f07fb6e801"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.702168 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68a4ac77-95df-4b11-958c-08f07fb6e801","Type":"ContainerDied","Data":"0c1dd37feb63b1a4d1ea39fd037b39ab6d62385a82820c82602f16566326fa01"} Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.702306 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.702288 4992 scope.go:117] "RemoveContainer" containerID="cc195017fa8e25b4716744641b10076a70498ff9ea29eae2b951f1edc63a791b" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.706799 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68a4ac77-95df-4b11-958c-08f07fb6e801" (UID: "68a4ac77-95df-4b11-958c-08f07fb6e801"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.764958 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7946\" (UniqueName: \"kubernetes.io/projected/68a4ac77-95df-4b11-958c-08f07fb6e801-kube-api-access-p7946\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.765350 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.765393 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68a4ac77-95df-4b11-958c-08f07fb6e801-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:43 crc kubenswrapper[4992]: I1004 04:55:43.776736 4992 scope.go:117] "RemoveContainer" containerID="6d705b5502e830807e782be561dd2cc95db110bc7563221f7f4e61bfd0acf329" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.045044 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.057793 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.075704 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:44 crc kubenswrapper[4992]: E1004 04:55:44.076286 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerName="nova-metadata-log" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.076307 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerName="nova-metadata-log" Oct 04 04:55:44 crc kubenswrapper[4992]: E1004 04:55:44.076481 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerName="nova-metadata-metadata" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.076493 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerName="nova-metadata-metadata" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.076706 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerName="nova-metadata-log" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.076731 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" containerName="nova-metadata-metadata" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.077824 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.080483 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.082159 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.096189 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.173088 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.173161 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd582411-fbae-4ed5-a39d-d3516cb95180-logs\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.173211 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rpz8\" (UniqueName: \"kubernetes.io/projected/fd582411-fbae-4ed5-a39d-d3516cb95180-kube-api-access-6rpz8\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.173236 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-config-data\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.173329 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.275210 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd582411-fbae-4ed5-a39d-d3516cb95180-logs\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.275623 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rpz8\" (UniqueName: \"kubernetes.io/projected/fd582411-fbae-4ed5-a39d-d3516cb95180-kube-api-access-6rpz8\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.275729 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-config-data\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.275874 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.275966 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.276213 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd582411-fbae-4ed5-a39d-d3516cb95180-logs\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.281129 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.281172 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.281822 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-config-data\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.293919 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rpz8\" (UniqueName: \"kubernetes.io/projected/fd582411-fbae-4ed5-a39d-d3516cb95180-kube-api-access-6rpz8\") pod \"nova-metadata-0\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.333836 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68a4ac77-95df-4b11-958c-08f07fb6e801" path="/var/lib/kubelet/pods/68a4ac77-95df-4b11-958c-08f07fb6e801/volumes" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.399810 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.709526 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.713655 4992 generic.go:334] "Generic (PLEG): container finished" podID="30b06dc6-6c8d-407b-bd1f-96f05c45660b" containerID="ebce72ef6dca56815743ca8f5b44109b59750eef660a14e596c39c5216e461ae" exitCode=0 Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.713689 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-b5b2w" event={"ID":"30b06dc6-6c8d-407b-bd1f-96f05c45660b","Type":"ContainerDied","Data":"ebce72ef6dca56815743ca8f5b44109b59750eef660a14e596c39c5216e461ae"} Oct 04 04:55:44 crc kubenswrapper[4992]: E1004 04:55:44.834609 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfffadcba_aa98_4d56_8b1f_86dc2435f8d0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b06dc6_6c8d_407b_bd1f_96f05c45660b.slice/crio-conmon-ebce72ef6dca56815743ca8f5b44109b59750eef660a14e596c39c5216e461ae.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.847924 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.847983 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.877957 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:44 crc kubenswrapper[4992]: W1004 04:55:44.878674 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd582411_fbae_4ed5_a39d_d3516cb95180.slice/crio-00219ca6134294902d553559b07302759b2a769edbf7c6673bbca322924b34b1 WatchSource:0}: Error finding container 00219ca6134294902d553559b07302759b2a769edbf7c6673bbca322924b34b1: Status 404 returned error can't find the container with id 00219ca6134294902d553559b07302759b2a769edbf7c6673bbca322924b34b1 Oct 04 04:55:44 crc kubenswrapper[4992]: I1004 04:55:44.921396 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.237868 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.239597 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.274179 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.301840 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.386006 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jzvwf"] Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.386433 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" podUID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerName="dnsmasq-dns" containerID="cri-o://c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84" gracePeriod=10 Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.727460 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd582411-fbae-4ed5-a39d-d3516cb95180","Type":"ContainerStarted","Data":"c85bf2dbd8e88677659071d20fb456a029c5a2eaca548d657b4aff000d3a7ba0"} Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.727908 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd582411-fbae-4ed5-a39d-d3516cb95180","Type":"ContainerStarted","Data":"00219ca6134294902d553559b07302759b2a769edbf7c6673bbca322924b34b1"} Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.773025 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.916252 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" podUID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.163:5353: connect: connection refused" Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.931709 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:55:45 crc kubenswrapper[4992]: I1004 04:55:45.931786 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.199917 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.318049 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-config-data\") pod \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.318137 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-scripts\") pod \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.318199 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-combined-ca-bundle\") pod \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.318296 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcgzn\" (UniqueName: \"kubernetes.io/projected/30b06dc6-6c8d-407b-bd1f-96f05c45660b-kube-api-access-wcgzn\") pod \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\" (UID: \"30b06dc6-6c8d-407b-bd1f-96f05c45660b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.324525 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b06dc6-6c8d-407b-bd1f-96f05c45660b-kube-api-access-wcgzn" (OuterVolumeSpecName: "kube-api-access-wcgzn") pod "30b06dc6-6c8d-407b-bd1f-96f05c45660b" (UID: "30b06dc6-6c8d-407b-bd1f-96f05c45660b"). InnerVolumeSpecName "kube-api-access-wcgzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.358584 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-scripts" (OuterVolumeSpecName: "scripts") pod "30b06dc6-6c8d-407b-bd1f-96f05c45660b" (UID: "30b06dc6-6c8d-407b-bd1f-96f05c45660b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.364748 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30b06dc6-6c8d-407b-bd1f-96f05c45660b" (UID: "30b06dc6-6c8d-407b-bd1f-96f05c45660b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.368179 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-config-data" (OuterVolumeSpecName: "config-data") pod "30b06dc6-6c8d-407b-bd1f-96f05c45660b" (UID: "30b06dc6-6c8d-407b-bd1f-96f05c45660b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.402613 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.429862 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.429898 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.429912 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcgzn\" (UniqueName: \"kubernetes.io/projected/30b06dc6-6c8d-407b-bd1f-96f05c45660b-kube-api-access-wcgzn\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.429925 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30b06dc6-6c8d-407b-bd1f-96f05c45660b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.531049 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nl2l\" (UniqueName: \"kubernetes.io/projected/808727d5-8ae0-4be5-ad58-b01ab7bff16b-kube-api-access-4nl2l\") pod \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.531156 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-sb\") pod \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.531214 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-svc\") pod \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.531284 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-config\") pod \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.531325 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-nb\") pod \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.531403 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-swift-storage-0\") pod \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\" (UID: \"808727d5-8ae0-4be5-ad58-b01ab7bff16b\") " Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.536568 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/808727d5-8ae0-4be5-ad58-b01ab7bff16b-kube-api-access-4nl2l" (OuterVolumeSpecName: "kube-api-access-4nl2l") pod "808727d5-8ae0-4be5-ad58-b01ab7bff16b" (UID: "808727d5-8ae0-4be5-ad58-b01ab7bff16b"). InnerVolumeSpecName "kube-api-access-4nl2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.585553 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "808727d5-8ae0-4be5-ad58-b01ab7bff16b" (UID: "808727d5-8ae0-4be5-ad58-b01ab7bff16b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.626473 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "808727d5-8ae0-4be5-ad58-b01ab7bff16b" (UID: "808727d5-8ae0-4be5-ad58-b01ab7bff16b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.626632 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-config" (OuterVolumeSpecName: "config") pod "808727d5-8ae0-4be5-ad58-b01ab7bff16b" (UID: "808727d5-8ae0-4be5-ad58-b01ab7bff16b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.634064 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "808727d5-8ae0-4be5-ad58-b01ab7bff16b" (UID: "808727d5-8ae0-4be5-ad58-b01ab7bff16b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.634382 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.634422 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.634436 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.634451 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.634463 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nl2l\" (UniqueName: \"kubernetes.io/projected/808727d5-8ae0-4be5-ad58-b01ab7bff16b-kube-api-access-4nl2l\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.640432 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "808727d5-8ae0-4be5-ad58-b01ab7bff16b" (UID: "808727d5-8ae0-4be5-ad58-b01ab7bff16b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.735537 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/808727d5-8ae0-4be5-ad58-b01ab7bff16b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.741765 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd582411-fbae-4ed5-a39d-d3516cb95180","Type":"ContainerStarted","Data":"b788e565deba0f52f04d1ee2524559b651833baa746c69956a065478a2970bda"} Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.746098 4992 generic.go:334] "Generic (PLEG): container finished" podID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerID="c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84" exitCode=0 Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.746229 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.746248 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" event={"ID":"808727d5-8ae0-4be5-ad58-b01ab7bff16b","Type":"ContainerDied","Data":"c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84"} Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.746279 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-jzvwf" event={"ID":"808727d5-8ae0-4be5-ad58-b01ab7bff16b","Type":"ContainerDied","Data":"1c74488a44757812d13b4a813739768401564eb189fc574a14be360198152dc8"} Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.746300 4992 scope.go:117] "RemoveContainer" containerID="c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.749549 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-b5b2w" event={"ID":"30b06dc6-6c8d-407b-bd1f-96f05c45660b","Type":"ContainerDied","Data":"d7f7d1c7c3a4c5081e29b1fd1811a408075748deaae1a5fdc47cf661b647bb1b"} Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.749591 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7f7d1c7c3a4c5081e29b1fd1811a408075748deaae1a5fdc47cf661b647bb1b" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.749748 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-b5b2w" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.769559 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.769538913 podStartE2EDuration="2.769538913s" podCreationTimestamp="2025-10-04 04:55:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:46.765809764 +0000 UTC m=+1380.613485232" watchObservedRunningTime="2025-10-04 04:55:46.769538913 +0000 UTC m=+1380.617214381" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.784787 4992 scope.go:117] "RemoveContainer" containerID="4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.808264 4992 scope.go:117] "RemoveContainer" containerID="c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.809388 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jzvwf"] Oct 04 04:55:46 crc kubenswrapper[4992]: E1004 04:55:46.809660 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84\": container with ID starting with c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84 not found: ID does not exist" containerID="c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.809702 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84"} err="failed to get container status \"c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84\": rpc error: code = NotFound desc = could not find container \"c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84\": container with ID starting with c772ffd4cba67492fff3336cae1c1fc2f7f2f5c369e5dcc03c946c4abfddec84 not found: ID does not exist" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.809732 4992 scope.go:117] "RemoveContainer" containerID="4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07" Oct 04 04:55:46 crc kubenswrapper[4992]: E1004 04:55:46.810171 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07\": container with ID starting with 4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07 not found: ID does not exist" containerID="4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.810201 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07"} err="failed to get container status \"4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07\": rpc error: code = NotFound desc = could not find container \"4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07\": container with ID starting with 4ed35740090ff401101c59d93aa5b6b079cc85f1a84f1798805e7d597263ac07 not found: ID does not exist" Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.818393 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-jzvwf"] Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.922418 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.922682 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-log" containerID="cri-o://e409510fc0911b8e2ead155ff70612039aac686fe8c37890cefc23a417435f61" gracePeriod=30 Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.922876 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-api" containerID="cri-o://a860dcaa514cd33c4368910e7b744e5e3a99aa0e7933bdafded92178606188ad" gracePeriod=30 Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.944689 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:46 crc kubenswrapper[4992]: I1004 04:55:46.958967 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:47 crc kubenswrapper[4992]: I1004 04:55:47.763168 4992 generic.go:334] "Generic (PLEG): container finished" podID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerID="e409510fc0911b8e2ead155ff70612039aac686fe8c37890cefc23a417435f61" exitCode=143 Oct 04 04:55:47 crc kubenswrapper[4992]: I1004 04:55:47.763268 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9","Type":"ContainerDied","Data":"e409510fc0911b8e2ead155ff70612039aac686fe8c37890cefc23a417435f61"} Oct 04 04:55:48 crc kubenswrapper[4992]: I1004 04:55:48.331243 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" path="/var/lib/kubelet/pods/808727d5-8ae0-4be5-ad58-b01ab7bff16b/volumes" Oct 04 04:55:48 crc kubenswrapper[4992]: I1004 04:55:48.576715 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:55:48 crc kubenswrapper[4992]: I1004 04:55:48.577233 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="ee9fdfcb-55b1-429c-bf7b-648b8d113c79" containerName="kube-state-metrics" containerID="cri-o://cab72e8d520cea797fd7ac44a178a47c14d70f36e382b4a489f6f665cd7b2c05" gracePeriod=30 Oct 04 04:55:48 crc kubenswrapper[4992]: I1004 04:55:48.773634 4992 generic.go:334] "Generic (PLEG): container finished" podID="ee9fdfcb-55b1-429c-bf7b-648b8d113c79" containerID="cab72e8d520cea797fd7ac44a178a47c14d70f36e382b4a489f6f665cd7b2c05" exitCode=2 Oct 04 04:55:48 crc kubenswrapper[4992]: I1004 04:55:48.773714 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ee9fdfcb-55b1-429c-bf7b-648b8d113c79","Type":"ContainerDied","Data":"cab72e8d520cea797fd7ac44a178a47c14d70f36e382b4a489f6f665cd7b2c05"} Oct 04 04:55:48 crc kubenswrapper[4992]: I1004 04:55:48.773841 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1fb7d1b9-f64a-4e58-92d7-f455273a7963" containerName="nova-scheduler-scheduler" containerID="cri-o://20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9" gracePeriod=30 Oct 04 04:55:48 crc kubenswrapper[4992]: I1004 04:55:48.773948 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerName="nova-metadata-log" containerID="cri-o://c85bf2dbd8e88677659071d20fb456a029c5a2eaca548d657b4aff000d3a7ba0" gracePeriod=30 Oct 04 04:55:48 crc kubenswrapper[4992]: I1004 04:55:48.774059 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerName="nova-metadata-metadata" containerID="cri-o://b788e565deba0f52f04d1ee2524559b651833baa746c69956a065478a2970bda" gracePeriod=30 Oct 04 04:55:49 crc kubenswrapper[4992]: I1004 04:55:49.400947 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:55:49 crc kubenswrapper[4992]: I1004 04:55:49.402896 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:55:49 crc kubenswrapper[4992]: I1004 04:55:49.789174 4992 generic.go:334] "Generic (PLEG): container finished" podID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerID="b788e565deba0f52f04d1ee2524559b651833baa746c69956a065478a2970bda" exitCode=0 Oct 04 04:55:49 crc kubenswrapper[4992]: I1004 04:55:49.789638 4992 generic.go:334] "Generic (PLEG): container finished" podID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerID="c85bf2dbd8e88677659071d20fb456a029c5a2eaca548d657b4aff000d3a7ba0" exitCode=143 Oct 04 04:55:49 crc kubenswrapper[4992]: I1004 04:55:49.789226 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd582411-fbae-4ed5-a39d-d3516cb95180","Type":"ContainerDied","Data":"b788e565deba0f52f04d1ee2524559b651833baa746c69956a065478a2970bda"} Oct 04 04:55:49 crc kubenswrapper[4992]: I1004 04:55:49.789680 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd582411-fbae-4ed5-a39d-d3516cb95180","Type":"ContainerDied","Data":"c85bf2dbd8e88677659071d20fb456a029c5a2eaca548d657b4aff000d3a7ba0"} Oct 04 04:55:49 crc kubenswrapper[4992]: I1004 04:55:49.936332 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:49.999320 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncv8f\" (UniqueName: \"kubernetes.io/projected/ee9fdfcb-55b1-429c-bf7b-648b8d113c79-kube-api-access-ncv8f\") pod \"ee9fdfcb-55b1-429c-bf7b-648b8d113c79\" (UID: \"ee9fdfcb-55b1-429c-bf7b-648b8d113c79\") " Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.006760 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9fdfcb-55b1-429c-bf7b-648b8d113c79-kube-api-access-ncv8f" (OuterVolumeSpecName: "kube-api-access-ncv8f") pod "ee9fdfcb-55b1-429c-bf7b-648b8d113c79" (UID: "ee9fdfcb-55b1-429c-bf7b-648b8d113c79"). InnerVolumeSpecName "kube-api-access-ncv8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.101118 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncv8f\" (UniqueName: \"kubernetes.io/projected/ee9fdfcb-55b1-429c-bf7b-648b8d113c79-kube-api-access-ncv8f\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.155919 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.239024 4992 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9 is running failed: container process not found" containerID="20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.239552 4992 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9 is running failed: container process not found" containerID="20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.240013 4992 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9 is running failed: container process not found" containerID="20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.240112 4992 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1fb7d1b9-f64a-4e58-92d7-f455273a7963" containerName="nova-scheduler-scheduler" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.278453 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.278813 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="ceilometer-central-agent" containerID="cri-o://51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318" gracePeriod=30 Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.278863 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="proxy-httpd" containerID="cri-o://f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318" gracePeriod=30 Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.278895 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="sg-core" containerID="cri-o://406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c" gracePeriod=30 Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.278910 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="ceilometer-notification-agent" containerID="cri-o://ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8" gracePeriod=30 Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.304119 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-nova-metadata-tls-certs\") pod \"fd582411-fbae-4ed5-a39d-d3516cb95180\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.304260 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-config-data\") pod \"fd582411-fbae-4ed5-a39d-d3516cb95180\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.304409 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rpz8\" (UniqueName: \"kubernetes.io/projected/fd582411-fbae-4ed5-a39d-d3516cb95180-kube-api-access-6rpz8\") pod \"fd582411-fbae-4ed5-a39d-d3516cb95180\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.304466 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd582411-fbae-4ed5-a39d-d3516cb95180-logs\") pod \"fd582411-fbae-4ed5-a39d-d3516cb95180\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.304493 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-combined-ca-bundle\") pod \"fd582411-fbae-4ed5-a39d-d3516cb95180\" (UID: \"fd582411-fbae-4ed5-a39d-d3516cb95180\") " Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.306242 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd582411-fbae-4ed5-a39d-d3516cb95180-logs" (OuterVolumeSpecName: "logs") pod "fd582411-fbae-4ed5-a39d-d3516cb95180" (UID: "fd582411-fbae-4ed5-a39d-d3516cb95180"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.309621 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd582411-fbae-4ed5-a39d-d3516cb95180-kube-api-access-6rpz8" (OuterVolumeSpecName: "kube-api-access-6rpz8") pod "fd582411-fbae-4ed5-a39d-d3516cb95180" (UID: "fd582411-fbae-4ed5-a39d-d3516cb95180"). InnerVolumeSpecName "kube-api-access-6rpz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.334597 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-config-data" (OuterVolumeSpecName: "config-data") pod "fd582411-fbae-4ed5-a39d-d3516cb95180" (UID: "fd582411-fbae-4ed5-a39d-d3516cb95180"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.336442 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd582411-fbae-4ed5-a39d-d3516cb95180" (UID: "fd582411-fbae-4ed5-a39d-d3516cb95180"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.356439 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fd582411-fbae-4ed5-a39d-d3516cb95180" (UID: "fd582411-fbae-4ed5-a39d-d3516cb95180"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.406689 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd582411-fbae-4ed5-a39d-d3516cb95180-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.406728 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.406741 4992 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.406752 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd582411-fbae-4ed5-a39d-d3516cb95180-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.406763 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rpz8\" (UniqueName: \"kubernetes.io/projected/fd582411-fbae-4ed5-a39d-d3516cb95180-kube-api-access-6rpz8\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.801080 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.801065 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd582411-fbae-4ed5-a39d-d3516cb95180","Type":"ContainerDied","Data":"00219ca6134294902d553559b07302759b2a769edbf7c6673bbca322924b34b1"} Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.801241 4992 scope.go:117] "RemoveContainer" containerID="b788e565deba0f52f04d1ee2524559b651833baa746c69956a065478a2970bda" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.804936 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ee9fdfcb-55b1-429c-bf7b-648b8d113c79","Type":"ContainerDied","Data":"3e8024c4952b7937aa08d356ab51d3b60f23c8fbf0955f6ee5d152ae9a249f97"} Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.804971 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.826735 4992 scope.go:117] "RemoveContainer" containerID="c85bf2dbd8e88677659071d20fb456a029c5a2eaca548d657b4aff000d3a7ba0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.835426 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.842470 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.854816 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.861251 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.861786 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerName="init" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.861814 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerName="init" Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.861832 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerName="nova-metadata-log" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.861841 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerName="nova-metadata-log" Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.861853 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerName="dnsmasq-dns" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.861862 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerName="dnsmasq-dns" Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.861913 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerName="nova-metadata-metadata" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.861921 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerName="nova-metadata-metadata" Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.861933 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9fdfcb-55b1-429c-bf7b-648b8d113c79" containerName="kube-state-metrics" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.861940 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9fdfcb-55b1-429c-bf7b-648b8d113c79" containerName="kube-state-metrics" Oct 04 04:55:50 crc kubenswrapper[4992]: E1004 04:55:50.861957 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b06dc6-6c8d-407b-bd1f-96f05c45660b" containerName="nova-manage" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.861964 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b06dc6-6c8d-407b-bd1f-96f05c45660b" containerName="nova-manage" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.862222 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="808727d5-8ae0-4be5-ad58-b01ab7bff16b" containerName="dnsmasq-dns" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.862245 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerName="nova-metadata-log" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.862257 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" containerName="nova-metadata-metadata" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.862274 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b06dc6-6c8d-407b-bd1f-96f05c45660b" containerName="nova-manage" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.862289 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9fdfcb-55b1-429c-bf7b-648b8d113c79" containerName="kube-state-metrics" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.863106 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.865716 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.866939 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.869772 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.880892 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.895526 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.897403 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.905155 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.905822 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.907821 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.914946 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.915101 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5ght\" (UniqueName: \"kubernetes.io/projected/f056bac1-081a-4675-865b-67f6a6035b47-kube-api-access-c5ght\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.915172 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.915244 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:50 crc kubenswrapper[4992]: I1004 04:55:50.926795 4992 scope.go:117] "RemoveContainer" containerID="cab72e8d520cea797fd7ac44a178a47c14d70f36e382b4a489f6f665cd7b2c05" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.017125 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.017196 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-config-data\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.017250 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3cc7753-b742-406f-abc1-b4b1c87ec984-logs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.017281 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.017333 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkgzs\" (UniqueName: \"kubernetes.io/projected/b3cc7753-b742-406f-abc1-b4b1c87ec984-kube-api-access-bkgzs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.025047 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.025174 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.025384 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.025407 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5ght\" (UniqueName: \"kubernetes.io/projected/f056bac1-081a-4675-865b-67f6a6035b47-kube-api-access-c5ght\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.028340 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.028652 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.035588 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f056bac1-081a-4675-865b-67f6a6035b47-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.046210 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5ght\" (UniqueName: \"kubernetes.io/projected/f056bac1-081a-4675-865b-67f6a6035b47-kube-api-access-c5ght\") pod \"kube-state-metrics-0\" (UID: \"f056bac1-081a-4675-865b-67f6a6035b47\") " pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.129570 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.129640 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-config-data\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.129678 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3cc7753-b742-406f-abc1-b4b1c87ec984-logs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.129709 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkgzs\" (UniqueName: \"kubernetes.io/projected/b3cc7753-b742-406f-abc1-b4b1c87ec984-kube-api-access-bkgzs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.129745 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.131563 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3cc7753-b742-406f-abc1-b4b1c87ec984-logs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.135530 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-config-data\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.136668 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.137669 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.149237 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkgzs\" (UniqueName: \"kubernetes.io/projected/b3cc7753-b742-406f-abc1-b4b1c87ec984-kube-api-access-bkgzs\") pod \"nova-metadata-0\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.209162 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.218033 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.760696 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:55:51 crc kubenswrapper[4992]: W1004 04:55:51.779380 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3cc7753_b742_406f_abc1_b4b1c87ec984.slice/crio-6f46b0684e9cc5e6073045d54c3a81234cc161df25adf0b29a74de2e0c5d9c5c WatchSource:0}: Error finding container 6f46b0684e9cc5e6073045d54c3a81234cc161df25adf0b29a74de2e0c5d9c5c: Status 404 returned error can't find the container with id 6f46b0684e9cc5e6073045d54c3a81234cc161df25adf0b29a74de2e0c5d9c5c Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.817135 4992 generic.go:334] "Generic (PLEG): container finished" podID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerID="f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318" exitCode=0 Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.817175 4992 generic.go:334] "Generic (PLEG): container finished" podID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerID="406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c" exitCode=2 Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.817185 4992 generic.go:334] "Generic (PLEG): container finished" podID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerID="51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318" exitCode=0 Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.817237 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerDied","Data":"f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318"} Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.817263 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerDied","Data":"406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c"} Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.817276 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerDied","Data":"51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318"} Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.818700 4992 generic.go:334] "Generic (PLEG): container finished" podID="1fb7d1b9-f64a-4e58-92d7-f455273a7963" containerID="20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9" exitCode=0 Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.818744 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1fb7d1b9-f64a-4e58-92d7-f455273a7963","Type":"ContainerDied","Data":"20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9"} Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.820219 4992 generic.go:334] "Generic (PLEG): container finished" podID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerID="a860dcaa514cd33c4368910e7b744e5e3a99aa0e7933bdafded92178606188ad" exitCode=0 Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.820267 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9","Type":"ContainerDied","Data":"a860dcaa514cd33c4368910e7b744e5e3a99aa0e7933bdafded92178606188ad"} Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.832130 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cc7753-b742-406f-abc1-b4b1c87ec984","Type":"ContainerStarted","Data":"6f46b0684e9cc5e6073045d54c3a81234cc161df25adf0b29a74de2e0c5d9c5c"} Oct 04 04:55:51 crc kubenswrapper[4992]: I1004 04:55:51.851791 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.351007 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9fdfcb-55b1-429c-bf7b-648b8d113c79" path="/var/lib/kubelet/pods/ee9fdfcb-55b1-429c-bf7b-648b8d113c79/volumes" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.351978 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd582411-fbae-4ed5-a39d-d3516cb95180" path="/var/lib/kubelet/pods/fd582411-fbae-4ed5-a39d-d3516cb95180/volumes" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.383773 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.446907 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.454490 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhdvn\" (UniqueName: \"kubernetes.io/projected/1fb7d1b9-f64a-4e58-92d7-f455273a7963-kube-api-access-dhdvn\") pod \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.454587 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle\") pod \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.454689 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-config-data\") pod \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.462370 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb7d1b9-f64a-4e58-92d7-f455273a7963-kube-api-access-dhdvn" (OuterVolumeSpecName: "kube-api-access-dhdvn") pod "1fb7d1b9-f64a-4e58-92d7-f455273a7963" (UID: "1fb7d1b9-f64a-4e58-92d7-f455273a7963"). InnerVolumeSpecName "kube-api-access-dhdvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4992]: E1004 04:55:52.485628 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle podName:1fb7d1b9-f64a-4e58-92d7-f455273a7963 nodeName:}" failed. No retries permitted until 2025-10-04 04:55:52.985599976 +0000 UTC m=+1386.833275444 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle") pod "1fb7d1b9-f64a-4e58-92d7-f455273a7963" (UID: "1fb7d1b9-f64a-4e58-92d7-f455273a7963") : error deleting /var/lib/kubelet/pods/1fb7d1b9-f64a-4e58-92d7-f455273a7963/volume-subpaths: remove /var/lib/kubelet/pods/1fb7d1b9-f64a-4e58-92d7-f455273a7963/volume-subpaths: no such file or directory Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.488591 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-config-data" (OuterVolumeSpecName: "config-data") pod "1fb7d1b9-f64a-4e58-92d7-f455273a7963" (UID: "1fb7d1b9-f64a-4e58-92d7-f455273a7963"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.556740 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-combined-ca-bundle\") pod \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.556805 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-logs\") pod \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.556955 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-config-data\") pod \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.557049 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb4x2\" (UniqueName: \"kubernetes.io/projected/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-kube-api-access-qb4x2\") pod \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\" (UID: \"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9\") " Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.557537 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhdvn\" (UniqueName: \"kubernetes.io/projected/1fb7d1b9-f64a-4e58-92d7-f455273a7963-kube-api-access-dhdvn\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.557557 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.557544 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-logs" (OuterVolumeSpecName: "logs") pod "70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" (UID: "70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.560115 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-kube-api-access-qb4x2" (OuterVolumeSpecName: "kube-api-access-qb4x2") pod "70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" (UID: "70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9"). InnerVolumeSpecName "kube-api-access-qb4x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.589879 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-config-data" (OuterVolumeSpecName: "config-data") pod "70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" (UID: "70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.592337 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" (UID: "70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.659672 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb4x2\" (UniqueName: \"kubernetes.io/projected/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-kube-api-access-qb4x2\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.660112 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.660231 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.660310 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.862080 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1fb7d1b9-f64a-4e58-92d7-f455273a7963","Type":"ContainerDied","Data":"d2af1ecfce7356592fc00a246b56df8e1c3341711f0546f33db368c25a836480"} Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.862093 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.863185 4992 scope.go:117] "RemoveContainer" containerID="20b7ce8656edd9914d80b6d4a78817604ff05b802647bc79b96e1786d355e4b9" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.866830 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.866996 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9","Type":"ContainerDied","Data":"fa04cee4feeb7543e483b79022a285ff78d0accf87ae1a7d281ef93d6009aaea"} Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.868748 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cc7753-b742-406f-abc1-b4b1c87ec984","Type":"ContainerStarted","Data":"6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d"} Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.868893 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cc7753-b742-406f-abc1-b4b1c87ec984","Type":"ContainerStarted","Data":"50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8"} Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.871242 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f056bac1-081a-4675-865b-67f6a6035b47","Type":"ContainerStarted","Data":"7639854f6705c2e679a533621a075cac6ebee0520e92150244632cc264e9dd5a"} Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.905913 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.905890908 podStartE2EDuration="2.905890908s" podCreationTimestamp="2025-10-04 04:55:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:52.893651372 +0000 UTC m=+1386.741326860" watchObservedRunningTime="2025-10-04 04:55:52.905890908 +0000 UTC m=+1386.753566376" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.922179 4992 scope.go:117] "RemoveContainer" containerID="a860dcaa514cd33c4368910e7b744e5e3a99aa0e7933bdafded92178606188ad" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.925531 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.944371 4992 scope.go:117] "RemoveContainer" containerID="e409510fc0911b8e2ead155ff70612039aac686fe8c37890cefc23a417435f61" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.944488 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.961104 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:52 crc kubenswrapper[4992]: E1004 04:55:52.961589 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb7d1b9-f64a-4e58-92d7-f455273a7963" containerName="nova-scheduler-scheduler" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.961614 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb7d1b9-f64a-4e58-92d7-f455273a7963" containerName="nova-scheduler-scheduler" Oct 04 04:55:52 crc kubenswrapper[4992]: E1004 04:55:52.961636 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-log" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.961643 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-log" Oct 04 04:55:52 crc kubenswrapper[4992]: E1004 04:55:52.961658 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-api" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.961664 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-api" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.961838 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-log" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.962053 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb7d1b9-f64a-4e58-92d7-f455273a7963" containerName="nova-scheduler-scheduler" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.962065 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" containerName="nova-api-api" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.963014 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.969142 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:55:52 crc kubenswrapper[4992]: I1004 04:55:52.969234 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.065986 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle\") pod \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\" (UID: \"1fb7d1b9-f64a-4e58-92d7-f455273a7963\") " Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.066382 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776930db-69ff-4dbb-95bf-2d81bf0b9b33-logs\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.066483 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhbgr\" (UniqueName: \"kubernetes.io/projected/776930db-69ff-4dbb-95bf-2d81bf0b9b33-kube-api-access-vhbgr\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.066553 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.066675 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-config-data\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.069988 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1fb7d1b9-f64a-4e58-92d7-f455273a7963" (UID: "1fb7d1b9-f64a-4e58-92d7-f455273a7963"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.168946 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhbgr\" (UniqueName: \"kubernetes.io/projected/776930db-69ff-4dbb-95bf-2d81bf0b9b33-kube-api-access-vhbgr\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.169022 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.169101 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-config-data\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.169177 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776930db-69ff-4dbb-95bf-2d81bf0b9b33-logs\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.169246 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fb7d1b9-f64a-4e58-92d7-f455273a7963-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.169614 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776930db-69ff-4dbb-95bf-2d81bf0b9b33-logs\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.173459 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-config-data\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.173934 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.187105 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhbgr\" (UniqueName: \"kubernetes.io/projected/776930db-69ff-4dbb-95bf-2d81bf0b9b33-kube-api-access-vhbgr\") pod \"nova-api-0\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.283814 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.328867 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.340748 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.350398 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.356952 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.358962 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.359551 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.475401 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-config-data\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.475553 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c8j4\" (UniqueName: \"kubernetes.io/projected/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-kube-api-access-7c8j4\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.475582 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.577065 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-config-data\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.577182 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c8j4\" (UniqueName: \"kubernetes.io/projected/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-kube-api-access-7c8j4\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.577210 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.581954 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-config-data\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.582003 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.602085 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c8j4\" (UniqueName: \"kubernetes.io/projected/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-kube-api-access-7c8j4\") pod \"nova-scheduler-0\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.680728 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.804182 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.885866 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-config-data\") pod \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.886009 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-scripts\") pod \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.888765 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-run-httpd\") pod \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.888819 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-log-httpd\") pod \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.888958 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-sg-core-conf-yaml\") pod \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.888995 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-combined-ca-bundle\") pod \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.889054 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmhc2\" (UniqueName: \"kubernetes.io/projected/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-kube-api-access-kmhc2\") pod \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\" (UID: \"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320\") " Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.890137 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" (UID: "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.890622 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-scripts" (OuterVolumeSpecName: "scripts") pod "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" (UID: "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.890742 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" (UID: "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.896464 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-kube-api-access-kmhc2" (OuterVolumeSpecName: "kube-api-access-kmhc2") pod "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" (UID: "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320"). InnerVolumeSpecName "kube-api-access-kmhc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.898301 4992 generic.go:334] "Generic (PLEG): container finished" podID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerID="ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8" exitCode=0 Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.898381 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerDied","Data":"ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8"} Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.898409 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eec65c5e-a9d7-42ef-9d2a-0787ab1d7320","Type":"ContainerDied","Data":"ae457492d1efcac55e6b2e86e8943b74adb9e5a8da05a53581a00b130ba57667"} Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.898424 4992 scope.go:117] "RemoveContainer" containerID="f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.898593 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.939492 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" (UID: "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.997641 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.997677 4992 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.997691 4992 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.997703 4992 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:53 crc kubenswrapper[4992]: I1004 04:55:53.997716 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmhc2\" (UniqueName: \"kubernetes.io/projected/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-kube-api-access-kmhc2\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.022934 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" (UID: "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.069263 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-config-data" (OuterVolumeSpecName: "config-data") pod "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" (UID: "eec65c5e-a9d7-42ef-9d2a-0787ab1d7320"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.099416 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.099617 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.105521 4992 scope.go:117] "RemoveContainer" containerID="406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.123676 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.148795 4992 scope.go:117] "RemoveContainer" containerID="ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.195337 4992 scope.go:117] "RemoveContainer" containerID="51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.217896 4992 scope.go:117] "RemoveContainer" containerID="f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318" Oct 04 04:55:54 crc kubenswrapper[4992]: E1004 04:55:54.218379 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318\": container with ID starting with f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318 not found: ID does not exist" containerID="f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.218420 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318"} err="failed to get container status \"f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318\": rpc error: code = NotFound desc = could not find container \"f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318\": container with ID starting with f42912f5414e5e48f6d9d15be74c423b305f32436edf7be42fde4bf988b65318 not found: ID does not exist" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.218452 4992 scope.go:117] "RemoveContainer" containerID="406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c" Oct 04 04:55:54 crc kubenswrapper[4992]: E1004 04:55:54.218802 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c\": container with ID starting with 406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c not found: ID does not exist" containerID="406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.218857 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c"} err="failed to get container status \"406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c\": rpc error: code = NotFound desc = could not find container \"406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c\": container with ID starting with 406dc2cfc8a8ae7b3475c3d4304c0848307ca788cca9ff2eb8d0f1ac4f0ae72c not found: ID does not exist" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.218890 4992 scope.go:117] "RemoveContainer" containerID="ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8" Oct 04 04:55:54 crc kubenswrapper[4992]: E1004 04:55:54.219440 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8\": container with ID starting with ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8 not found: ID does not exist" containerID="ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.219470 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8"} err="failed to get container status \"ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8\": rpc error: code = NotFound desc = could not find container \"ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8\": container with ID starting with ee3727179b9845192c4274fda6f0f18999a968f9a679111058d855bd313c87b8 not found: ID does not exist" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.219489 4992 scope.go:117] "RemoveContainer" containerID="51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318" Oct 04 04:55:54 crc kubenswrapper[4992]: E1004 04:55:54.221870 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318\": container with ID starting with 51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318 not found: ID does not exist" containerID="51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.221895 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318"} err="failed to get container status \"51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318\": rpc error: code = NotFound desc = could not find container \"51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318\": container with ID starting with 51a0afbe10c313300a5854c73089eaf4573d20ed210345388706ee9a447f7318 not found: ID does not exist" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.250675 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.272386 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.288694 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:54 crc kubenswrapper[4992]: E1004 04:55:54.289232 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="proxy-httpd" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.289258 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="proxy-httpd" Oct 04 04:55:54 crc kubenswrapper[4992]: E1004 04:55:54.289282 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="ceilometer-notification-agent" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.289298 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="ceilometer-notification-agent" Oct 04 04:55:54 crc kubenswrapper[4992]: E1004 04:55:54.289338 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="ceilometer-central-agent" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.289349 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="ceilometer-central-agent" Oct 04 04:55:54 crc kubenswrapper[4992]: E1004 04:55:54.289426 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="sg-core" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.289436 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="sg-core" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.289708 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="proxy-httpd" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.289741 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="ceilometer-central-agent" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.289774 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="sg-core" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.289799 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" containerName="ceilometer-notification-agent" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.291998 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.294924 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.295076 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.296931 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.298616 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.330537 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb7d1b9-f64a-4e58-92d7-f455273a7963" path="/var/lib/kubelet/pods/1fb7d1b9-f64a-4e58-92d7-f455273a7963/volumes" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.331081 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9" path="/var/lib/kubelet/pods/70502243-4cc6-4f13-ae6e-b1f5c8ad8fb9/volumes" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.331710 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eec65c5e-a9d7-42ef-9d2a-0787ab1d7320" path="/var/lib/kubelet/pods/eec65c5e-a9d7-42ef-9d2a-0787ab1d7320/volumes" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.333092 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.407163 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-log-httpd\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.407239 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-scripts\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.407286 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpfq9\" (UniqueName: \"kubernetes.io/projected/6bc9e190-4719-4192-983c-03c428f1ed84-kube-api-access-wpfq9\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.407311 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.407385 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-config-data\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.407412 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-run-httpd\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.407429 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.407453 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.509271 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-run-httpd\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.509316 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.509345 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.509384 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-log-httpd\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.509456 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-scripts\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.510496 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-run-httpd\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.510620 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-log-httpd\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.510719 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpfq9\" (UniqueName: \"kubernetes.io/projected/6bc9e190-4719-4192-983c-03c428f1ed84-kube-api-access-wpfq9\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.510771 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.511152 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-config-data\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.514517 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.515145 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-config-data\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.515483 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.516727 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-scripts\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.521869 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.526673 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpfq9\" (UniqueName: \"kubernetes.io/projected/6bc9e190-4719-4192-983c-03c428f1ed84-kube-api-access-wpfq9\") pod \"ceilometer-0\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.626122 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.926889 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37690cca-04c7-41d4-8ffa-36c9bf09c9b6","Type":"ContainerStarted","Data":"72584ab7b963251881167fb87357ffd30791de1deb7ec50da7e5ea3e1cfaa650"} Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.927257 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37690cca-04c7-41d4-8ffa-36c9bf09c9b6","Type":"ContainerStarted","Data":"a42e64b2a35ac02336e345b3e18bfc4485c76ffbb0d990baacd830d2377eddb3"} Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.929141 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f056bac1-081a-4675-865b-67f6a6035b47","Type":"ContainerStarted","Data":"6519b98924eb6de9808f89f7f75971a212cfd2670018256d81942ca25034fbfb"} Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.929459 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.931802 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"776930db-69ff-4dbb-95bf-2d81bf0b9b33","Type":"ContainerStarted","Data":"8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233"} Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.931836 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"776930db-69ff-4dbb-95bf-2d81bf0b9b33","Type":"ContainerStarted","Data":"d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95"} Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.931846 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"776930db-69ff-4dbb-95bf-2d81bf0b9b33","Type":"ContainerStarted","Data":"01d96206f11b78cbc0807053bffd9627aefd03b80ba6bde72c6cb8c75b817cf8"} Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.945018 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.94499724 podStartE2EDuration="1.94499724s" podCreationTimestamp="2025-10-04 04:55:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:54.94422083 +0000 UTC m=+1388.791896308" watchObservedRunningTime="2025-10-04 04:55:54.94499724 +0000 UTC m=+1388.792672708" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.963340 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.159999211 podStartE2EDuration="4.963324827s" podCreationTimestamp="2025-10-04 04:55:50 +0000 UTC" firstStartedPulling="2025-10-04 04:55:51.858899706 +0000 UTC m=+1385.706575174" lastFinishedPulling="2025-10-04 04:55:53.662225282 +0000 UTC m=+1387.509900790" observedRunningTime="2025-10-04 04:55:54.962195527 +0000 UTC m=+1388.809871015" watchObservedRunningTime="2025-10-04 04:55:54.963324827 +0000 UTC m=+1388.811000285" Oct 04 04:55:54 crc kubenswrapper[4992]: I1004 04:55:54.992904 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.992882063 podStartE2EDuration="2.992882063s" podCreationTimestamp="2025-10-04 04:55:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:55:54.985210849 +0000 UTC m=+1388.832886327" watchObservedRunningTime="2025-10-04 04:55:54.992882063 +0000 UTC m=+1388.840557531" Oct 04 04:55:55 crc kubenswrapper[4992]: I1004 04:55:55.152618 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:55:55 crc kubenswrapper[4992]: I1004 04:55:55.948026 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerStarted","Data":"0e30b44a60a227c962011099c653c08b9f3a76a1abdcb1720ea7d151b71d27f3"} Oct 04 04:55:56 crc kubenswrapper[4992]: I1004 04:55:56.218134 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:55:56 crc kubenswrapper[4992]: I1004 04:55:56.218418 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:55:56 crc kubenswrapper[4992]: I1004 04:55:56.958044 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerStarted","Data":"77515990fc208dfe3a4ac0806b9984f73fe631843eec21b537591e8db1dffa51"} Oct 04 04:55:56 crc kubenswrapper[4992]: I1004 04:55:56.958293 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerStarted","Data":"21bfee8250581a79c116d16342f7b1fccf51ec429e7927d8deea72270c7f5193"} Oct 04 04:55:57 crc kubenswrapper[4992]: I1004 04:55:57.970096 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerStarted","Data":"a39692110716d57b5deebbbe8249739a6c1141c43fdd884a7dba8e7a956b4114"} Oct 04 04:55:58 crc kubenswrapper[4992]: I1004 04:55:58.681382 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:55:58 crc kubenswrapper[4992]: I1004 04:55:58.994810 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerStarted","Data":"c077359219ffa989eb527dd2ee5349d788e958b6de39cc45c31f6caa98e8963e"} Oct 04 04:55:58 crc kubenswrapper[4992]: I1004 04:55:58.996529 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:55:59 crc kubenswrapper[4992]: I1004 04:55:59.026630 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.812609632 podStartE2EDuration="5.026607796s" podCreationTimestamp="2025-10-04 04:55:54 +0000 UTC" firstStartedPulling="2025-10-04 04:55:55.156851822 +0000 UTC m=+1389.004527290" lastFinishedPulling="2025-10-04 04:55:58.370849986 +0000 UTC m=+1392.218525454" observedRunningTime="2025-10-04 04:55:59.016412395 +0000 UTC m=+1392.864087883" watchObservedRunningTime="2025-10-04 04:55:59.026607796 +0000 UTC m=+1392.874283264" Oct 04 04:56:01 crc kubenswrapper[4992]: I1004 04:56:01.018568 4992 generic.go:334] "Generic (PLEG): container finished" podID="2cc135c0-8c87-4802-836a-8c59def0eb0d" containerID="1ceb7c240fb308cad875d20402abe1d34f87bdc00eb054ac85c8762e2ae05835" exitCode=0 Oct 04 04:56:01 crc kubenswrapper[4992]: I1004 04:56:01.018661 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" event={"ID":"2cc135c0-8c87-4802-836a-8c59def0eb0d","Type":"ContainerDied","Data":"1ceb7c240fb308cad875d20402abe1d34f87bdc00eb054ac85c8762e2ae05835"} Oct 04 04:56:01 crc kubenswrapper[4992]: I1004 04:56:01.218258 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:56:01 crc kubenswrapper[4992]: I1004 04:56:01.219472 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:56:01 crc kubenswrapper[4992]: I1004 04:56:01.220340 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.229573 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.229618 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.465398 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.587847 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-scripts\") pod \"2cc135c0-8c87-4802-836a-8c59def0eb0d\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.588043 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-combined-ca-bundle\") pod \"2cc135c0-8c87-4802-836a-8c59def0eb0d\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.588080 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6pgt\" (UniqueName: \"kubernetes.io/projected/2cc135c0-8c87-4802-836a-8c59def0eb0d-kube-api-access-j6pgt\") pod \"2cc135c0-8c87-4802-836a-8c59def0eb0d\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.588110 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-config-data\") pod \"2cc135c0-8c87-4802-836a-8c59def0eb0d\" (UID: \"2cc135c0-8c87-4802-836a-8c59def0eb0d\") " Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.593975 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-scripts" (OuterVolumeSpecName: "scripts") pod "2cc135c0-8c87-4802-836a-8c59def0eb0d" (UID: "2cc135c0-8c87-4802-836a-8c59def0eb0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.598346 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cc135c0-8c87-4802-836a-8c59def0eb0d-kube-api-access-j6pgt" (OuterVolumeSpecName: "kube-api-access-j6pgt") pod "2cc135c0-8c87-4802-836a-8c59def0eb0d" (UID: "2cc135c0-8c87-4802-836a-8c59def0eb0d"). InnerVolumeSpecName "kube-api-access-j6pgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.620396 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-config-data" (OuterVolumeSpecName: "config-data") pod "2cc135c0-8c87-4802-836a-8c59def0eb0d" (UID: "2cc135c0-8c87-4802-836a-8c59def0eb0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.626689 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2cc135c0-8c87-4802-836a-8c59def0eb0d" (UID: "2cc135c0-8c87-4802-836a-8c59def0eb0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.690431 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.690478 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6pgt\" (UniqueName: \"kubernetes.io/projected/2cc135c0-8c87-4802-836a-8c59def0eb0d-kube-api-access-j6pgt\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.690496 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:02 crc kubenswrapper[4992]: I1004 04:56:02.690507 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cc135c0-8c87-4802-836a-8c59def0eb0d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.040833 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" event={"ID":"2cc135c0-8c87-4802-836a-8c59def0eb0d","Type":"ContainerDied","Data":"c0474efb7e7260850b9a088445c8e58a89f5087e9ff363ce534de2bd603ecf4a"} Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.040902 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0474efb7e7260850b9a088445c8e58a89f5087e9ff363ce534de2bd603ecf4a" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.040967 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-ts9f9" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.118137 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:56:03 crc kubenswrapper[4992]: E1004 04:56:03.118522 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cc135c0-8c87-4802-836a-8c59def0eb0d" containerName="nova-cell1-conductor-db-sync" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.118538 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cc135c0-8c87-4802-836a-8c59def0eb0d" containerName="nova-cell1-conductor-db-sync" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.118728 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cc135c0-8c87-4802-836a-8c59def0eb0d" containerName="nova-cell1-conductor-db-sync" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.119335 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.121814 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.135272 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.200578 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gd6v\" (UniqueName: \"kubernetes.io/projected/f291828a-08bf-400b-95d0-af8e6e7ecd10-kube-api-access-4gd6v\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.201038 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f291828a-08bf-400b-95d0-af8e6e7ecd10-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.201208 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f291828a-08bf-400b-95d0-af8e6e7ecd10-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.284516 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.284561 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.302980 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f291828a-08bf-400b-95d0-af8e6e7ecd10-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.303085 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f291828a-08bf-400b-95d0-af8e6e7ecd10-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.303167 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gd6v\" (UniqueName: \"kubernetes.io/projected/f291828a-08bf-400b-95d0-af8e6e7ecd10-kube-api-access-4gd6v\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.309148 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f291828a-08bf-400b-95d0-af8e6e7ecd10-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.318285 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f291828a-08bf-400b-95d0-af8e6e7ecd10-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.325946 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gd6v\" (UniqueName: \"kubernetes.io/projected/f291828a-08bf-400b-95d0-af8e6e7ecd10-kube-api-access-4gd6v\") pod \"nova-cell1-conductor-0\" (UID: \"f291828a-08bf-400b-95d0-af8e6e7ecd10\") " pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.441608 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.687934 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.737819 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:56:03 crc kubenswrapper[4992]: I1004 04:56:03.785420 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 04:56:03 crc kubenswrapper[4992]: W1004 04:56:03.789102 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf291828a_08bf_400b_95d0_af8e6e7ecd10.slice/crio-c5fa2733f0d5ea57b91dada6ae27f7f111d7edf3e5593efee93e8d4e44d25afd WatchSource:0}: Error finding container c5fa2733f0d5ea57b91dada6ae27f7f111d7edf3e5593efee93e8d4e44d25afd: Status 404 returned error can't find the container with id c5fa2733f0d5ea57b91dada6ae27f7f111d7edf3e5593efee93e8d4e44d25afd Oct 04 04:56:04 crc kubenswrapper[4992]: I1004 04:56:04.051196 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f291828a-08bf-400b-95d0-af8e6e7ecd10","Type":"ContainerStarted","Data":"c5fa2733f0d5ea57b91dada6ae27f7f111d7edf3e5593efee93e8d4e44d25afd"} Oct 04 04:56:04 crc kubenswrapper[4992]: I1004 04:56:04.079116 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:56:04 crc kubenswrapper[4992]: I1004 04:56:04.368650 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:04 crc kubenswrapper[4992]: I1004 04:56:04.368654 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:05 crc kubenswrapper[4992]: I1004 04:56:05.065695 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f291828a-08bf-400b-95d0-af8e6e7ecd10","Type":"ContainerStarted","Data":"37fbea898ab6d34c01cd63ff0538eb29ac94922ba8b295db0d3afac26e75e01d"} Oct 04 04:56:05 crc kubenswrapper[4992]: I1004 04:56:05.091306 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.091289246 podStartE2EDuration="2.091289246s" podCreationTimestamp="2025-10-04 04:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:05.086875559 +0000 UTC m=+1398.934551027" watchObservedRunningTime="2025-10-04 04:56:05.091289246 +0000 UTC m=+1398.938964714" Oct 04 04:56:06 crc kubenswrapper[4992]: I1004 04:56:06.073124 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.122784 4992 generic.go:334] "Generic (PLEG): container finished" podID="9497ef49-7631-453f-8c40-2fe991a823fa" containerID="2a2358f1c9f9d51ebe58d267edcafd4b0425263be03c3045909384bceca080a5" exitCode=137 Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.122891 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9497ef49-7631-453f-8c40-2fe991a823fa","Type":"ContainerDied","Data":"2a2358f1c9f9d51ebe58d267edcafd4b0425263be03c3045909384bceca080a5"} Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.123209 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9497ef49-7631-453f-8c40-2fe991a823fa","Type":"ContainerDied","Data":"924d5c8a46b9f6d1ad16d196c55255e3f35259edb2f2db8c282a5010597e0ae4"} Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.123221 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="924d5c8a46b9f6d1ad16d196c55255e3f35259edb2f2db8c282a5010597e0ae4" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.184928 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.223467 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.227114 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.228007 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.264352 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-combined-ca-bundle\") pod \"9497ef49-7631-453f-8c40-2fe991a823fa\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.264415 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjw2c\" (UniqueName: \"kubernetes.io/projected/9497ef49-7631-453f-8c40-2fe991a823fa-kube-api-access-qjw2c\") pod \"9497ef49-7631-453f-8c40-2fe991a823fa\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.264471 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-config-data\") pod \"9497ef49-7631-453f-8c40-2fe991a823fa\" (UID: \"9497ef49-7631-453f-8c40-2fe991a823fa\") " Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.269552 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9497ef49-7631-453f-8c40-2fe991a823fa-kube-api-access-qjw2c" (OuterVolumeSpecName: "kube-api-access-qjw2c") pod "9497ef49-7631-453f-8c40-2fe991a823fa" (UID: "9497ef49-7631-453f-8c40-2fe991a823fa"). InnerVolumeSpecName "kube-api-access-qjw2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.304169 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9497ef49-7631-453f-8c40-2fe991a823fa" (UID: "9497ef49-7631-453f-8c40-2fe991a823fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.308953 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-config-data" (OuterVolumeSpecName: "config-data") pod "9497ef49-7631-453f-8c40-2fe991a823fa" (UID: "9497ef49-7631-453f-8c40-2fe991a823fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.368235 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.368302 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjw2c\" (UniqueName: \"kubernetes.io/projected/9497ef49-7631-453f-8c40-2fe991a823fa-kube-api-access-qjw2c\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:11 crc kubenswrapper[4992]: I1004 04:56:11.368318 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9497ef49-7631-453f-8c40-2fe991a823fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.131037 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.136208 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.175837 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.183980 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.208741 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:56:12 crc kubenswrapper[4992]: E1004 04:56:12.209711 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9497ef49-7631-453f-8c40-2fe991a823fa" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.209736 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9497ef49-7631-453f-8c40-2fe991a823fa" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.210003 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9497ef49-7631-453f-8c40-2fe991a823fa" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.210775 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.213294 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.213524 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.213659 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.246845 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.285677 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.285736 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.285913 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.285969 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.286063 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw5bw\" (UniqueName: \"kubernetes.io/projected/1544458b-29c9-465f-adc2-f5ad257303fc-kube-api-access-xw5bw\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.329601 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9497ef49-7631-453f-8c40-2fe991a823fa" path="/var/lib/kubelet/pods/9497ef49-7631-453f-8c40-2fe991a823fa/volumes" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.387403 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.387487 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.387572 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.387611 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.387657 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw5bw\" (UniqueName: \"kubernetes.io/projected/1544458b-29c9-465f-adc2-f5ad257303fc-kube-api-access-xw5bw\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.392003 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.395079 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.395559 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.403789 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1544458b-29c9-465f-adc2-f5ad257303fc-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.404908 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw5bw\" (UniqueName: \"kubernetes.io/projected/1544458b-29c9-465f-adc2-f5ad257303fc-kube-api-access-xw5bw\") pod \"nova-cell1-novncproxy-0\" (UID: \"1544458b-29c9-465f-adc2-f5ad257303fc\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:12 crc kubenswrapper[4992]: I1004 04:56:12.537884 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:13 crc kubenswrapper[4992]: I1004 04:56:13.020454 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 04:56:13 crc kubenswrapper[4992]: I1004 04:56:13.141495 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1544458b-29c9-465f-adc2-f5ad257303fc","Type":"ContainerStarted","Data":"f69264ceba8fdd5d52fa7e9991d9c8ce48e6e3069c5ce5deb3e2e4f167d261a4"} Oct 04 04:56:13 crc kubenswrapper[4992]: I1004 04:56:13.287755 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:56:13 crc kubenswrapper[4992]: I1004 04:56:13.288514 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:56:13 crc kubenswrapper[4992]: I1004 04:56:13.288869 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:56:13 crc kubenswrapper[4992]: I1004 04:56:13.292204 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:56:13 crc kubenswrapper[4992]: I1004 04:56:13.474094 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.150797 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1544458b-29c9-465f-adc2-f5ad257303fc","Type":"ContainerStarted","Data":"9fff329dca81340b9de5e70ccd56e19f96eae90653ae412ea462a0f4ba7e53b9"} Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.151083 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.156174 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.171894 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.171874913 podStartE2EDuration="2.171874913s" podCreationTimestamp="2025-10-04 04:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:14.166245693 +0000 UTC m=+1408.013921161" watchObservedRunningTime="2025-10-04 04:56:14.171874913 +0000 UTC m=+1408.019550371" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.374492 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-th48l"] Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.390557 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.418470 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-th48l"] Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.429851 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drr24\" (UniqueName: \"kubernetes.io/projected/30b4e051-7811-4273-a749-629ec4cff7a8-kube-api-access-drr24\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.429902 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-config\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.429971 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.430000 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.430020 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.430052 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.532122 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.532188 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.532211 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.532239 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.532431 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drr24\" (UniqueName: \"kubernetes.io/projected/30b4e051-7811-4273-a749-629ec4cff7a8-kube-api-access-drr24\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.532464 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-config\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.533301 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.533507 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.535303 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.535792 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-config\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.538124 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.557344 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drr24\" (UniqueName: \"kubernetes.io/projected/30b4e051-7811-4273-a749-629ec4cff7a8-kube-api-access-drr24\") pod \"dnsmasq-dns-59cf4bdb65-th48l\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:14 crc kubenswrapper[4992]: I1004 04:56:14.721647 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:15 crc kubenswrapper[4992]: I1004 04:56:15.218942 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-th48l"] Oct 04 04:56:15 crc kubenswrapper[4992]: W1004 04:56:15.222378 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b4e051_7811_4273_a749_629ec4cff7a8.slice/crio-9d134cdeeb17ebb0b7b672a4a2c5ad347afc47e9902d86a5da10a320e9b78a28 WatchSource:0}: Error finding container 9d134cdeeb17ebb0b7b672a4a2c5ad347afc47e9902d86a5da10a320e9b78a28: Status 404 returned error can't find the container with id 9d134cdeeb17ebb0b7b672a4a2c5ad347afc47e9902d86a5da10a320e9b78a28 Oct 04 04:56:15 crc kubenswrapper[4992]: E1004 04:56:15.672713 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b4e051_7811_4273_a749_629ec4cff7a8.slice/crio-conmon-dc9119f28a854c2cf653ff0929786845541ee6e0bad370943f4adb60e304a67d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b4e051_7811_4273_a749_629ec4cff7a8.slice/crio-dc9119f28a854c2cf653ff0929786845541ee6e0bad370943f4adb60e304a67d.scope\": RecentStats: unable to find data in memory cache]" Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.168945 4992 generic.go:334] "Generic (PLEG): container finished" podID="30b4e051-7811-4273-a749-629ec4cff7a8" containerID="dc9119f28a854c2cf653ff0929786845541ee6e0bad370943f4adb60e304a67d" exitCode=0 Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.168998 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" event={"ID":"30b4e051-7811-4273-a749-629ec4cff7a8","Type":"ContainerDied","Data":"dc9119f28a854c2cf653ff0929786845541ee6e0bad370943f4adb60e304a67d"} Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.169282 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" event={"ID":"30b4e051-7811-4273-a749-629ec4cff7a8","Type":"ContainerStarted","Data":"9d134cdeeb17ebb0b7b672a4a2c5ad347afc47e9902d86a5da10a320e9b78a28"} Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.403151 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.403808 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="ceilometer-central-agent" containerID="cri-o://21bfee8250581a79c116d16342f7b1fccf51ec429e7927d8deea72270c7f5193" gracePeriod=30 Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.404515 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="proxy-httpd" containerID="cri-o://c077359219ffa989eb527dd2ee5349d788e958b6de39cc45c31f6caa98e8963e" gracePeriod=30 Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.404558 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="sg-core" containerID="cri-o://a39692110716d57b5deebbbe8249739a6c1141c43fdd884a7dba8e7a956b4114" gracePeriod=30 Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.404551 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="ceilometer-notification-agent" containerID="cri-o://77515990fc208dfe3a4ac0806b9984f73fe631843eec21b537591e8db1dffa51" gracePeriod=30 Oct 04 04:56:16 crc kubenswrapper[4992]: I1004 04:56:16.412685 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.196:3000/\": read tcp 10.217.0.2:40390->10.217.0.196:3000: read: connection reset by peer" Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.185127 4992 generic.go:334] "Generic (PLEG): container finished" podID="6bc9e190-4719-4192-983c-03c428f1ed84" containerID="c077359219ffa989eb527dd2ee5349d788e958b6de39cc45c31f6caa98e8963e" exitCode=0 Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.185162 4992 generic.go:334] "Generic (PLEG): container finished" podID="6bc9e190-4719-4192-983c-03c428f1ed84" containerID="a39692110716d57b5deebbbe8249739a6c1141c43fdd884a7dba8e7a956b4114" exitCode=2 Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.185172 4992 generic.go:334] "Generic (PLEG): container finished" podID="6bc9e190-4719-4192-983c-03c428f1ed84" containerID="21bfee8250581a79c116d16342f7b1fccf51ec429e7927d8deea72270c7f5193" exitCode=0 Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.185196 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerDied","Data":"c077359219ffa989eb527dd2ee5349d788e958b6de39cc45c31f6caa98e8963e"} Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.185247 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerDied","Data":"a39692110716d57b5deebbbe8249739a6c1141c43fdd884a7dba8e7a956b4114"} Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.185262 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerDied","Data":"21bfee8250581a79c116d16342f7b1fccf51ec429e7927d8deea72270c7f5193"} Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.187278 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" event={"ID":"30b4e051-7811-4273-a749-629ec4cff7a8","Type":"ContainerStarted","Data":"78b97b081cb35ea7a4e1436df09d9b89c2ba5cada56bf249e43048a561fe3eab"} Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.187561 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.189213 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.189394 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-log" containerID="cri-o://d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95" gracePeriod=30 Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.189458 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-api" containerID="cri-o://8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233" gracePeriod=30 Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.213329 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" podStartSLOduration=3.213306259 podStartE2EDuration="3.213306259s" podCreationTimestamp="2025-10-04 04:56:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:17.209452237 +0000 UTC m=+1411.057127715" watchObservedRunningTime="2025-10-04 04:56:17.213306259 +0000 UTC m=+1411.060981727" Oct 04 04:56:17 crc kubenswrapper[4992]: I1004 04:56:17.538707 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:18 crc kubenswrapper[4992]: I1004 04:56:18.197096 4992 generic.go:334] "Generic (PLEG): container finished" podID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerID="d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95" exitCode=143 Oct 04 04:56:18 crc kubenswrapper[4992]: I1004 04:56:18.197188 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"776930db-69ff-4dbb-95bf-2d81bf0b9b33","Type":"ContainerDied","Data":"d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95"} Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.233884 4992 generic.go:334] "Generic (PLEG): container finished" podID="6bc9e190-4719-4192-983c-03c428f1ed84" containerID="77515990fc208dfe3a4ac0806b9984f73fe631843eec21b537591e8db1dffa51" exitCode=0 Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.233963 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerDied","Data":"77515990fc208dfe3a4ac0806b9984f73fe631843eec21b537591e8db1dffa51"} Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.234389 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6bc9e190-4719-4192-983c-03c428f1ed84","Type":"ContainerDied","Data":"0e30b44a60a227c962011099c653c08b9f3a76a1abdcb1720ea7d151b71d27f3"} Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.234409 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e30b44a60a227c962011099c653c08b9f3a76a1abdcb1720ea7d151b71d27f3" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.285744 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.455022 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-sg-core-conf-yaml\") pod \"6bc9e190-4719-4192-983c-03c428f1ed84\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.455075 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-ceilometer-tls-certs\") pod \"6bc9e190-4719-4192-983c-03c428f1ed84\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.455185 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-run-httpd\") pod \"6bc9e190-4719-4192-983c-03c428f1ed84\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.455239 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-scripts\") pod \"6bc9e190-4719-4192-983c-03c428f1ed84\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.455327 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpfq9\" (UniqueName: \"kubernetes.io/projected/6bc9e190-4719-4192-983c-03c428f1ed84-kube-api-access-wpfq9\") pod \"6bc9e190-4719-4192-983c-03c428f1ed84\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.455409 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-combined-ca-bundle\") pod \"6bc9e190-4719-4192-983c-03c428f1ed84\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.455425 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-config-data\") pod \"6bc9e190-4719-4192-983c-03c428f1ed84\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.455466 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-log-httpd\") pod \"6bc9e190-4719-4192-983c-03c428f1ed84\" (UID: \"6bc9e190-4719-4192-983c-03c428f1ed84\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.456387 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6bc9e190-4719-4192-983c-03c428f1ed84" (UID: "6bc9e190-4719-4192-983c-03c428f1ed84"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.457849 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6bc9e190-4719-4192-983c-03c428f1ed84" (UID: "6bc9e190-4719-4192-983c-03c428f1ed84"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.496736 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bc9e190-4719-4192-983c-03c428f1ed84-kube-api-access-wpfq9" (OuterVolumeSpecName: "kube-api-access-wpfq9") pod "6bc9e190-4719-4192-983c-03c428f1ed84" (UID: "6bc9e190-4719-4192-983c-03c428f1ed84"). InnerVolumeSpecName "kube-api-access-wpfq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.497854 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-scripts" (OuterVolumeSpecName: "scripts") pod "6bc9e190-4719-4192-983c-03c428f1ed84" (UID: "6bc9e190-4719-4192-983c-03c428f1ed84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.521636 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6bc9e190-4719-4192-983c-03c428f1ed84" (UID: "6bc9e190-4719-4192-983c-03c428f1ed84"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.524145 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "6bc9e190-4719-4192-983c-03c428f1ed84" (UID: "6bc9e190-4719-4192-983c-03c428f1ed84"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.557540 4992 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.557814 4992 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.557828 4992 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.557840 4992 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6bc9e190-4719-4192-983c-03c428f1ed84-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.557852 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.557864 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpfq9\" (UniqueName: \"kubernetes.io/projected/6bc9e190-4719-4192-983c-03c428f1ed84-kube-api-access-wpfq9\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.559428 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bc9e190-4719-4192-983c-03c428f1ed84" (UID: "6bc9e190-4719-4192-983c-03c428f1ed84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.599400 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-config-data" (OuterVolumeSpecName: "config-data") pod "6bc9e190-4719-4192-983c-03c428f1ed84" (UID: "6bc9e190-4719-4192-983c-03c428f1ed84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.659819 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.659877 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bc9e190-4719-4192-983c-03c428f1ed84-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.731861 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.863978 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-combined-ca-bundle\") pod \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.864482 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-config-data\") pod \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.864643 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhbgr\" (UniqueName: \"kubernetes.io/projected/776930db-69ff-4dbb-95bf-2d81bf0b9b33-kube-api-access-vhbgr\") pod \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.864699 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776930db-69ff-4dbb-95bf-2d81bf0b9b33-logs\") pod \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\" (UID: \"776930db-69ff-4dbb-95bf-2d81bf0b9b33\") " Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.865880 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/776930db-69ff-4dbb-95bf-2d81bf0b9b33-logs" (OuterVolumeSpecName: "logs") pod "776930db-69ff-4dbb-95bf-2d81bf0b9b33" (UID: "776930db-69ff-4dbb-95bf-2d81bf0b9b33"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.868078 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/776930db-69ff-4dbb-95bf-2d81bf0b9b33-kube-api-access-vhbgr" (OuterVolumeSpecName: "kube-api-access-vhbgr") pod "776930db-69ff-4dbb-95bf-2d81bf0b9b33" (UID: "776930db-69ff-4dbb-95bf-2d81bf0b9b33"). InnerVolumeSpecName "kube-api-access-vhbgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.896484 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "776930db-69ff-4dbb-95bf-2d81bf0b9b33" (UID: "776930db-69ff-4dbb-95bf-2d81bf0b9b33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.899114 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-config-data" (OuterVolumeSpecName: "config-data") pod "776930db-69ff-4dbb-95bf-2d81bf0b9b33" (UID: "776930db-69ff-4dbb-95bf-2d81bf0b9b33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.968161 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.968224 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776930db-69ff-4dbb-95bf-2d81bf0b9b33-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.968238 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhbgr\" (UniqueName: \"kubernetes.io/projected/776930db-69ff-4dbb-95bf-2d81bf0b9b33-kube-api-access-vhbgr\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:20 crc kubenswrapper[4992]: I1004 04:56:20.968276 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776930db-69ff-4dbb-95bf-2d81bf0b9b33-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.244841 4992 generic.go:334] "Generic (PLEG): container finished" podID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerID="8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233" exitCode=0 Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.244885 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"776930db-69ff-4dbb-95bf-2d81bf0b9b33","Type":"ContainerDied","Data":"8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233"} Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.244926 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.244932 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.244954 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"776930db-69ff-4dbb-95bf-2d81bf0b9b33","Type":"ContainerDied","Data":"01d96206f11b78cbc0807053bffd9627aefd03b80ba6bde72c6cb8c75b817cf8"} Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.244977 4992 scope.go:117] "RemoveContainer" containerID="8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.266151 4992 scope.go:117] "RemoveContainer" containerID="d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.287696 4992 scope.go:117] "RemoveContainer" containerID="8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233" Oct 04 04:56:21 crc kubenswrapper[4992]: E1004 04:56:21.288055 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233\": container with ID starting with 8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233 not found: ID does not exist" containerID="8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.288083 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233"} err="failed to get container status \"8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233\": rpc error: code = NotFound desc = could not find container \"8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233\": container with ID starting with 8116dfcc1b977901ae37a70ec13b6895dc354962b0c01bc36f87240167a47233 not found: ID does not exist" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.288108 4992 scope.go:117] "RemoveContainer" containerID="d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95" Oct 04 04:56:21 crc kubenswrapper[4992]: E1004 04:56:21.288300 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95\": container with ID starting with d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95 not found: ID does not exist" containerID="d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.288321 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95"} err="failed to get container status \"d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95\": rpc error: code = NotFound desc = could not find container \"d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95\": container with ID starting with d644ce7e82b5b0060545599c136f1d3c863b958c84ad600c632e2060a628da95 not found: ID does not exist" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.289755 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.313672 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.328792 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.337455 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.349190 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:56:21 crc kubenswrapper[4992]: E1004 04:56:21.349703 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="sg-core" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.349723 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="sg-core" Oct 04 04:56:21 crc kubenswrapper[4992]: E1004 04:56:21.349738 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="ceilometer-notification-agent" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.349745 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="ceilometer-notification-agent" Oct 04 04:56:21 crc kubenswrapper[4992]: E1004 04:56:21.349762 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-api" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.349769 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-api" Oct 04 04:56:21 crc kubenswrapper[4992]: E1004 04:56:21.349791 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-log" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.349797 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-log" Oct 04 04:56:21 crc kubenswrapper[4992]: E1004 04:56:21.349808 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="ceilometer-central-agent" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.349814 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="ceilometer-central-agent" Oct 04 04:56:21 crc kubenswrapper[4992]: E1004 04:56:21.349833 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="proxy-httpd" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.349841 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="proxy-httpd" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.350051 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-api" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.350071 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="sg-core" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.350089 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="proxy-httpd" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.350104 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="ceilometer-central-agent" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.350116 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" containerName="ceilometer-notification-agent" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.350129 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" containerName="nova-api-log" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.352306 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.355088 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.355679 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.355765 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.385946 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.407055 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.409008 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.411113 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.411502 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.412618 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.413066 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.479559 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-scripts\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.480301 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.480556 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc693b9c-b61e-4794-b736-59f588ee80a1-run-httpd\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.481018 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v747j\" (UniqueName: \"kubernetes.io/projected/bc693b9c-b61e-4794-b736-59f588ee80a1-kube-api-access-v747j\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.481402 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.481466 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-config-data\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.481673 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.481794 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc693b9c-b61e-4794-b736-59f588ee80a1-log-httpd\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584239 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584305 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-config-data\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584341 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-config-data\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584420 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584448 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc693b9c-b61e-4794-b736-59f588ee80a1-log-httpd\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584477 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584507 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-scripts\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584583 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584613 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584638 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc693b9c-b61e-4794-b736-59f588ee80a1-run-httpd\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584672 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v747j\" (UniqueName: \"kubernetes.io/projected/bc693b9c-b61e-4794-b736-59f588ee80a1-kube-api-access-v747j\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584694 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmrzc\" (UniqueName: \"kubernetes.io/projected/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-kube-api-access-lmrzc\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584776 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-logs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.584818 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-public-tls-certs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.585837 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc693b9c-b61e-4794-b736-59f588ee80a1-run-httpd\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.585952 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc693b9c-b61e-4794-b736-59f588ee80a1-log-httpd\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.588762 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-scripts\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.592825 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.592882 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.593139 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-config-data\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.599885 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc693b9c-b61e-4794-b736-59f588ee80a1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.601082 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v747j\" (UniqueName: \"kubernetes.io/projected/bc693b9c-b61e-4794-b736-59f588ee80a1-kube-api-access-v747j\") pod \"ceilometer-0\" (UID: \"bc693b9c-b61e-4794-b736-59f588ee80a1\") " pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.685394 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.686237 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-logs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.686317 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-public-tls-certs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.686408 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-config-data\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.686476 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.686561 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.686595 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmrzc\" (UniqueName: \"kubernetes.io/projected/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-kube-api-access-lmrzc\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.688774 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-logs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.690996 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-config-data\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.691028 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.692265 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-public-tls-certs\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.695972 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.712400 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmrzc\" (UniqueName: \"kubernetes.io/projected/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-kube-api-access-lmrzc\") pod \"nova-api-0\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " pod="openstack/nova-api-0" Oct 04 04:56:21 crc kubenswrapper[4992]: I1004 04:56:21.730275 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:56:22 crc kubenswrapper[4992]: I1004 04:56:22.152559 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 04:56:22 crc kubenswrapper[4992]: I1004 04:56:22.160976 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 04:56:22 crc kubenswrapper[4992]: I1004 04:56:22.240530 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:22 crc kubenswrapper[4992]: W1004 04:56:22.245150 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fca9cad_13e6_4ff2_a5c4_8c3e31133cfd.slice/crio-7d2e189d44abf91ff34f90073903348a997d506477bb36e6fcf15af390a9c0d1 WatchSource:0}: Error finding container 7d2e189d44abf91ff34f90073903348a997d506477bb36e6fcf15af390a9c0d1: Status 404 returned error can't find the container with id 7d2e189d44abf91ff34f90073903348a997d506477bb36e6fcf15af390a9c0d1 Oct 04 04:56:22 crc kubenswrapper[4992]: I1004 04:56:22.266912 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerStarted","Data":"67c6ad1d35955a0353505e7a7b5f1081943e6ddfecbad4eb8d9155eb7e8dc62d"} Oct 04 04:56:22 crc kubenswrapper[4992]: I1004 04:56:22.330886 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bc9e190-4719-4192-983c-03c428f1ed84" path="/var/lib/kubelet/pods/6bc9e190-4719-4192-983c-03c428f1ed84/volumes" Oct 04 04:56:22 crc kubenswrapper[4992]: I1004 04:56:22.332627 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="776930db-69ff-4dbb-95bf-2d81bf0b9b33" path="/var/lib/kubelet/pods/776930db-69ff-4dbb-95bf-2d81bf0b9b33/volumes" Oct 04 04:56:22 crc kubenswrapper[4992]: I1004 04:56:22.538316 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:22 crc kubenswrapper[4992]: I1004 04:56:22.559130 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.277470 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd","Type":"ContainerStarted","Data":"2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b"} Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.278791 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd","Type":"ContainerStarted","Data":"25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee"} Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.278906 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd","Type":"ContainerStarted","Data":"7d2e189d44abf91ff34f90073903348a997d506477bb36e6fcf15af390a9c0d1"} Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.279773 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerStarted","Data":"715c12561023a725df6254da792b9cc8bdcf7f098a7935409b169e6d19f5cf5b"} Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.294599 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.302433 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.302413789 podStartE2EDuration="2.302413789s" podCreationTimestamp="2025-10-04 04:56:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:23.297112818 +0000 UTC m=+1417.144788296" watchObservedRunningTime="2025-10-04 04:56:23.302413789 +0000 UTC m=+1417.150089257" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.444006 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-cwhkv"] Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.445448 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.449477 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.452942 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.454664 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-cwhkv"] Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.525250 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-scripts\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.525519 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-config-data\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.525719 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.525853 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvsz6\" (UniqueName: \"kubernetes.io/projected/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-kube-api-access-jvsz6\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.627695 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-config-data\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.628088 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.628183 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvsz6\" (UniqueName: \"kubernetes.io/projected/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-kube-api-access-jvsz6\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.628836 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-scripts\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.632444 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.632463 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-config-data\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.634882 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-scripts\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.649028 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvsz6\" (UniqueName: \"kubernetes.io/projected/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-kube-api-access-jvsz6\") pod \"nova-cell1-cell-mapping-cwhkv\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:23 crc kubenswrapper[4992]: I1004 04:56:23.766432 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:24 crc kubenswrapper[4992]: I1004 04:56:24.246340 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-cwhkv"] Oct 04 04:56:24 crc kubenswrapper[4992]: I1004 04:56:24.300849 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cwhkv" event={"ID":"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d","Type":"ContainerStarted","Data":"ffc6784363586661ad106f675d8d6821396b015d23e03c5c2c264e0901832bb0"} Oct 04 04:56:24 crc kubenswrapper[4992]: I1004 04:56:24.303513 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerStarted","Data":"99cafd3f4d325875ed8a631e1d05e044cf4ecd9bbab472ce70775ba5ee5c1aea"} Oct 04 04:56:24 crc kubenswrapper[4992]: I1004 04:56:24.723580 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:56:24 crc kubenswrapper[4992]: I1004 04:56:24.793395 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-8xvd8"] Oct 04 04:56:24 crc kubenswrapper[4992]: I1004 04:56:24.793829 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" podUID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerName="dnsmasq-dns" containerID="cri-o://472db47216833d39a4fb8cca0b5a138ac1238adcf1e32c55b306b33fb9eaf815" gracePeriod=10 Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.313768 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cwhkv" event={"ID":"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d","Type":"ContainerStarted","Data":"ed5b2ca893c80e74fd5b3ee930d3a75e23f161b9e1839ef31cbce3468343f7de"} Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.316564 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerStarted","Data":"3fea8d5bb1ed6397d8f4c8c2b6e74aedbfc81603d50357fadecc1770e8c6c7a6"} Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.319087 4992 generic.go:334] "Generic (PLEG): container finished" podID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerID="472db47216833d39a4fb8cca0b5a138ac1238adcf1e32c55b306b33fb9eaf815" exitCode=0 Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.319101 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" event={"ID":"7e30d629-2fa2-4a60-be4d-cc73893f5454","Type":"ContainerDied","Data":"472db47216833d39a4fb8cca0b5a138ac1238adcf1e32c55b306b33fb9eaf815"} Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.319239 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" event={"ID":"7e30d629-2fa2-4a60-be4d-cc73893f5454","Type":"ContainerDied","Data":"219c28379870f40df08115184b886814aaa17c59833571655b3abfb1fe7f0898"} Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.319254 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="219c28379870f40df08115184b886814aaa17c59833571655b3abfb1fe7f0898" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.338219 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.344292 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-cwhkv" podStartSLOduration=2.344271254 podStartE2EDuration="2.344271254s" podCreationTimestamp="2025-10-04 04:56:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:25.332018708 +0000 UTC m=+1419.179694216" watchObservedRunningTime="2025-10-04 04:56:25.344271254 +0000 UTC m=+1419.191946722" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.407268 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc5lt\" (UniqueName: \"kubernetes.io/projected/7e30d629-2fa2-4a60-be4d-cc73893f5454-kube-api-access-cc5lt\") pod \"7e30d629-2fa2-4a60-be4d-cc73893f5454\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.407329 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-sb\") pod \"7e30d629-2fa2-4a60-be4d-cc73893f5454\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.407535 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-nb\") pod \"7e30d629-2fa2-4a60-be4d-cc73893f5454\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.407622 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-config\") pod \"7e30d629-2fa2-4a60-be4d-cc73893f5454\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.407689 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-svc\") pod \"7e30d629-2fa2-4a60-be4d-cc73893f5454\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.407755 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-swift-storage-0\") pod \"7e30d629-2fa2-4a60-be4d-cc73893f5454\" (UID: \"7e30d629-2fa2-4a60-be4d-cc73893f5454\") " Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.415522 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e30d629-2fa2-4a60-be4d-cc73893f5454-kube-api-access-cc5lt" (OuterVolumeSpecName: "kube-api-access-cc5lt") pod "7e30d629-2fa2-4a60-be4d-cc73893f5454" (UID: "7e30d629-2fa2-4a60-be4d-cc73893f5454"). InnerVolumeSpecName "kube-api-access-cc5lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.462997 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e30d629-2fa2-4a60-be4d-cc73893f5454" (UID: "7e30d629-2fa2-4a60-be4d-cc73893f5454"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.471137 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e30d629-2fa2-4a60-be4d-cc73893f5454" (UID: "7e30d629-2fa2-4a60-be4d-cc73893f5454"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.471691 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7e30d629-2fa2-4a60-be4d-cc73893f5454" (UID: "7e30d629-2fa2-4a60-be4d-cc73893f5454"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.478132 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-config" (OuterVolumeSpecName: "config") pod "7e30d629-2fa2-4a60-be4d-cc73893f5454" (UID: "7e30d629-2fa2-4a60-be4d-cc73893f5454"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.478626 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e30d629-2fa2-4a60-be4d-cc73893f5454" (UID: "7e30d629-2fa2-4a60-be4d-cc73893f5454"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.510411 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.510454 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.510472 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc5lt\" (UniqueName: \"kubernetes.io/projected/7e30d629-2fa2-4a60-be4d-cc73893f5454-kube-api-access-cc5lt\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.510484 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.510495 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:25 crc kubenswrapper[4992]: I1004 04:56:25.510506 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e30d629-2fa2-4a60-be4d-cc73893f5454-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:26 crc kubenswrapper[4992]: I1004 04:56:26.342604 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" Oct 04 04:56:26 crc kubenswrapper[4992]: I1004 04:56:26.345759 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerStarted","Data":"2047ba60b5f16c1ba226b2a0a6dbe1a74eb3aa4918564f92d796cf9368bee2c0"} Oct 04 04:56:26 crc kubenswrapper[4992]: I1004 04:56:26.345818 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 04:56:26 crc kubenswrapper[4992]: I1004 04:56:26.386667 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.968672168 podStartE2EDuration="5.386647503s" podCreationTimestamp="2025-10-04 04:56:21 +0000 UTC" firstStartedPulling="2025-10-04 04:56:22.160789092 +0000 UTC m=+1416.008464560" lastFinishedPulling="2025-10-04 04:56:25.578764427 +0000 UTC m=+1419.426439895" observedRunningTime="2025-10-04 04:56:26.376104142 +0000 UTC m=+1420.223779610" watchObservedRunningTime="2025-10-04 04:56:26.386647503 +0000 UTC m=+1420.234322971" Oct 04 04:56:26 crc kubenswrapper[4992]: I1004 04:56:26.399381 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-8xvd8"] Oct 04 04:56:26 crc kubenswrapper[4992]: I1004 04:56:26.407064 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-8xvd8"] Oct 04 04:56:28 crc kubenswrapper[4992]: I1004 04:56:28.331042 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e30d629-2fa2-4a60-be4d-cc73893f5454" path="/var/lib/kubelet/pods/7e30d629-2fa2-4a60-be4d-cc73893f5454/volumes" Oct 04 04:56:30 crc kubenswrapper[4992]: I1004 04:56:30.300004 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-845d6d6f59-8xvd8" podUID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.189:5353: i/o timeout" Oct 04 04:56:30 crc kubenswrapper[4992]: I1004 04:56:30.386708 4992 generic.go:334] "Generic (PLEG): container finished" podID="1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" containerID="ed5b2ca893c80e74fd5b3ee930d3a75e23f161b9e1839ef31cbce3468343f7de" exitCode=0 Oct 04 04:56:30 crc kubenswrapper[4992]: I1004 04:56:30.386763 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cwhkv" event={"ID":"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d","Type":"ContainerDied","Data":"ed5b2ca893c80e74fd5b3ee930d3a75e23f161b9e1839ef31cbce3468343f7de"} Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.730808 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.731303 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.795534 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.950803 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvsz6\" (UniqueName: \"kubernetes.io/projected/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-kube-api-access-jvsz6\") pod \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.950913 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-combined-ca-bundle\") pod \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.951091 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-config-data\") pod \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.951129 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-scripts\") pod \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\" (UID: \"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d\") " Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.959301 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-kube-api-access-jvsz6" (OuterVolumeSpecName: "kube-api-access-jvsz6") pod "1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" (UID: "1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d"). InnerVolumeSpecName "kube-api-access-jvsz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.967538 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-scripts" (OuterVolumeSpecName: "scripts") pod "1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" (UID: "1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.980864 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-config-data" (OuterVolumeSpecName: "config-data") pod "1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" (UID: "1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:31 crc kubenswrapper[4992]: I1004 04:56:31.986616 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" (UID: "1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.053395 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.053774 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.053785 4992 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.053795 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvsz6\" (UniqueName: \"kubernetes.io/projected/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d-kube-api-access-jvsz6\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.442884 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cwhkv" event={"ID":"1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d","Type":"ContainerDied","Data":"ffc6784363586661ad106f675d8d6821396b015d23e03c5c2c264e0901832bb0"} Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.442958 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cwhkv" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.442974 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffc6784363586661ad106f675d8d6821396b015d23e03c5c2c264e0901832bb0" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.580107 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.580406 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-log" containerID="cri-o://25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee" gracePeriod=30 Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.580524 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-api" containerID="cri-o://2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b" gracePeriod=30 Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.584474 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": EOF" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.584647 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": EOF" Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.590227 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.590604 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="37690cca-04c7-41d4-8ffa-36c9bf09c9b6" containerName="nova-scheduler-scheduler" containerID="cri-o://72584ab7b963251881167fb87357ffd30791de1deb7ec50da7e5ea3e1cfaa650" gracePeriod=30 Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.610759 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.611123 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-log" containerID="cri-o://50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8" gracePeriod=30 Oct 04 04:56:32 crc kubenswrapper[4992]: I1004 04:56:32.611189 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-metadata" containerID="cri-o://6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d" gracePeriod=30 Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.463784 4992 generic.go:334] "Generic (PLEG): container finished" podID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerID="50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8" exitCode=143 Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.464066 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cc7753-b742-406f-abc1-b4b1c87ec984","Type":"ContainerDied","Data":"50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8"} Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.472236 4992 generic.go:334] "Generic (PLEG): container finished" podID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerID="25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee" exitCode=143 Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.472315 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd","Type":"ContainerDied","Data":"25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee"} Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.473572 4992 generic.go:334] "Generic (PLEG): container finished" podID="37690cca-04c7-41d4-8ffa-36c9bf09c9b6" containerID="72584ab7b963251881167fb87357ffd30791de1deb7ec50da7e5ea3e1cfaa650" exitCode=0 Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.473602 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37690cca-04c7-41d4-8ffa-36c9bf09c9b6","Type":"ContainerDied","Data":"72584ab7b963251881167fb87357ffd30791de1deb7ec50da7e5ea3e1cfaa650"} Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.473621 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37690cca-04c7-41d4-8ffa-36c9bf09c9b6","Type":"ContainerDied","Data":"a42e64b2a35ac02336e345b3e18bfc4485c76ffbb0d990baacd830d2377eddb3"} Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.473634 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a42e64b2a35ac02336e345b3e18bfc4485c76ffbb0d990baacd830d2377eddb3" Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.532106 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.697996 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-combined-ca-bundle\") pod \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.698066 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-config-data\") pod \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.698097 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c8j4\" (UniqueName: \"kubernetes.io/projected/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-kube-api-access-7c8j4\") pod \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\" (UID: \"37690cca-04c7-41d4-8ffa-36c9bf09c9b6\") " Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.703318 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-kube-api-access-7c8j4" (OuterVolumeSpecName: "kube-api-access-7c8j4") pod "37690cca-04c7-41d4-8ffa-36c9bf09c9b6" (UID: "37690cca-04c7-41d4-8ffa-36c9bf09c9b6"). InnerVolumeSpecName "kube-api-access-7c8j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.726638 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-config-data" (OuterVolumeSpecName: "config-data") pod "37690cca-04c7-41d4-8ffa-36c9bf09c9b6" (UID: "37690cca-04c7-41d4-8ffa-36c9bf09c9b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.734792 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37690cca-04c7-41d4-8ffa-36c9bf09c9b6" (UID: "37690cca-04c7-41d4-8ffa-36c9bf09c9b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.800262 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.800296 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:33 crc kubenswrapper[4992]: I1004 04:56:33.800305 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c8j4\" (UniqueName: \"kubernetes.io/projected/37690cca-04c7-41d4-8ffa-36c9bf09c9b6-kube-api-access-7c8j4\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.482517 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.503195 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.545259 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.568116 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:56:34 crc kubenswrapper[4992]: E1004 04:56:34.568617 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" containerName="nova-manage" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.568726 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" containerName="nova-manage" Oct 04 04:56:34 crc kubenswrapper[4992]: E1004 04:56:34.568756 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37690cca-04c7-41d4-8ffa-36c9bf09c9b6" containerName="nova-scheduler-scheduler" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.568765 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="37690cca-04c7-41d4-8ffa-36c9bf09c9b6" containerName="nova-scheduler-scheduler" Oct 04 04:56:34 crc kubenswrapper[4992]: E1004 04:56:34.568789 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerName="dnsmasq-dns" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.568798 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerName="dnsmasq-dns" Oct 04 04:56:34 crc kubenswrapper[4992]: E1004 04:56:34.568837 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerName="init" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.568845 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerName="init" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.569070 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" containerName="nova-manage" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.569094 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="37690cca-04c7-41d4-8ffa-36c9bf09c9b6" containerName="nova-scheduler-scheduler" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.569110 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e30d629-2fa2-4a60-be4d-cc73893f5454" containerName="dnsmasq-dns" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.569864 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.571789 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.577418 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.724243 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-config-data\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.724298 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.724354 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc7gh\" (UniqueName: \"kubernetes.io/projected/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-kube-api-access-gc7gh\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.825864 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-config-data\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.825939 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.826013 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc7gh\" (UniqueName: \"kubernetes.io/projected/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-kube-api-access-gc7gh\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.830961 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.831027 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-config-data\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.848024 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc7gh\" (UniqueName: \"kubernetes.io/projected/bcf9657e-a56d-48fa-906a-aeaa6f3d3672-kube-api-access-gc7gh\") pod \"nova-scheduler-0\" (UID: \"bcf9657e-a56d-48fa-906a-aeaa6f3d3672\") " pod="openstack/nova-scheduler-0" Oct 04 04:56:34 crc kubenswrapper[4992]: I1004 04:56:34.891859 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 04:56:35 crc kubenswrapper[4992]: I1004 04:56:35.321577 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 04:56:35 crc kubenswrapper[4992]: I1004 04:56:35.496651 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bcf9657e-a56d-48fa-906a-aeaa6f3d3672","Type":"ContainerStarted","Data":"08356feaeed7d4127d0821ceb811004ad79afdc55ec05f6e6a1c822ba94eba61"} Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.228122 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.332308 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37690cca-04c7-41d4-8ffa-36c9bf09c9b6" path="/var/lib/kubelet/pods/37690cca-04c7-41d4-8ffa-36c9bf09c9b6/volumes" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.352284 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-combined-ca-bundle\") pod \"b3cc7753-b742-406f-abc1-b4b1c87ec984\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.352388 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-nova-metadata-tls-certs\") pod \"b3cc7753-b742-406f-abc1-b4b1c87ec984\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.352479 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3cc7753-b742-406f-abc1-b4b1c87ec984-logs\") pod \"b3cc7753-b742-406f-abc1-b4b1c87ec984\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.352737 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-config-data\") pod \"b3cc7753-b742-406f-abc1-b4b1c87ec984\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.352837 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkgzs\" (UniqueName: \"kubernetes.io/projected/b3cc7753-b742-406f-abc1-b4b1c87ec984-kube-api-access-bkgzs\") pod \"b3cc7753-b742-406f-abc1-b4b1c87ec984\" (UID: \"b3cc7753-b742-406f-abc1-b4b1c87ec984\") " Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.353155 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3cc7753-b742-406f-abc1-b4b1c87ec984-logs" (OuterVolumeSpecName: "logs") pod "b3cc7753-b742-406f-abc1-b4b1c87ec984" (UID: "b3cc7753-b742-406f-abc1-b4b1c87ec984"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.356293 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3cc7753-b742-406f-abc1-b4b1c87ec984-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.358081 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3cc7753-b742-406f-abc1-b4b1c87ec984-kube-api-access-bkgzs" (OuterVolumeSpecName: "kube-api-access-bkgzs") pod "b3cc7753-b742-406f-abc1-b4b1c87ec984" (UID: "b3cc7753-b742-406f-abc1-b4b1c87ec984"). InnerVolumeSpecName "kube-api-access-bkgzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.402102 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3cc7753-b742-406f-abc1-b4b1c87ec984" (UID: "b3cc7753-b742-406f-abc1-b4b1c87ec984"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.402455 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-config-data" (OuterVolumeSpecName: "config-data") pod "b3cc7753-b742-406f-abc1-b4b1c87ec984" (UID: "b3cc7753-b742-406f-abc1-b4b1c87ec984"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.434420 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b3cc7753-b742-406f-abc1-b4b1c87ec984" (UID: "b3cc7753-b742-406f-abc1-b4b1c87ec984"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.459614 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.459657 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkgzs\" (UniqueName: \"kubernetes.io/projected/b3cc7753-b742-406f-abc1-b4b1c87ec984-kube-api-access-bkgzs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.459671 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.459686 4992 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b3cc7753-b742-406f-abc1-b4b1c87ec984-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.506476 4992 generic.go:334] "Generic (PLEG): container finished" podID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerID="6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d" exitCode=0 Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.506542 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.506548 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cc7753-b742-406f-abc1-b4b1c87ec984","Type":"ContainerDied","Data":"6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d"} Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.506788 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cc7753-b742-406f-abc1-b4b1c87ec984","Type":"ContainerDied","Data":"6f46b0684e9cc5e6073045d54c3a81234cc161df25adf0b29a74de2e0c5d9c5c"} Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.506840 4992 scope.go:117] "RemoveContainer" containerID="6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.510808 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bcf9657e-a56d-48fa-906a-aeaa6f3d3672","Type":"ContainerStarted","Data":"fa0c405aee5d7c98808cd5b4743e0ddffe2a39f3550aaba02884b04c490e7319"} Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.551627 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.551599967 podStartE2EDuration="2.551599967s" podCreationTimestamp="2025-10-04 04:56:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:36.530187719 +0000 UTC m=+1430.377863187" watchObservedRunningTime="2025-10-04 04:56:36.551599967 +0000 UTC m=+1430.399275435" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.558987 4992 scope.go:117] "RemoveContainer" containerID="50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.573518 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.584122 4992 scope.go:117] "RemoveContainer" containerID="6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d" Oct 04 04:56:36 crc kubenswrapper[4992]: E1004 04:56:36.584508 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d\": container with ID starting with 6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d not found: ID does not exist" containerID="6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.584551 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d"} err="failed to get container status \"6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d\": rpc error: code = NotFound desc = could not find container \"6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d\": container with ID starting with 6eed7cebe4e1d8aaa9f9d94ab063174f78ccda44344ff10ecf74db1288d1796d not found: ID does not exist" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.584578 4992 scope.go:117] "RemoveContainer" containerID="50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8" Oct 04 04:56:36 crc kubenswrapper[4992]: E1004 04:56:36.584844 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8\": container with ID starting with 50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8 not found: ID does not exist" containerID="50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.584872 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8"} err="failed to get container status \"50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8\": rpc error: code = NotFound desc = could not find container \"50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8\": container with ID starting with 50e27f981835c1b2b55cf8e258d8f74c8c175ef5954379e381424f13d2276ed8 not found: ID does not exist" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.591560 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.603945 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:56:36 crc kubenswrapper[4992]: E1004 04:56:36.604431 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-metadata" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.604455 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-metadata" Oct 04 04:56:36 crc kubenswrapper[4992]: E1004 04:56:36.604480 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-log" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.604489 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-log" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.604757 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-metadata" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.604782 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-log" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.605927 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.609789 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.610006 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.622387 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.764913 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.764986 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgx7t\" (UniqueName: \"kubernetes.io/projected/f399940f-164b-4033-af0a-36359efe7229-kube-api-access-bgx7t\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.765037 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f399940f-164b-4033-af0a-36359efe7229-logs\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.765210 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-config-data\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.765443 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.869687 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.869787 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgx7t\" (UniqueName: \"kubernetes.io/projected/f399940f-164b-4033-af0a-36359efe7229-kube-api-access-bgx7t\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.869825 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f399940f-164b-4033-af0a-36359efe7229-logs\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.869906 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-config-data\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.869984 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.870475 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f399940f-164b-4033-af0a-36359efe7229-logs\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.875040 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-config-data\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.875841 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.875965 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f399940f-164b-4033-af0a-36359efe7229-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.889165 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgx7t\" (UniqueName: \"kubernetes.io/projected/f399940f-164b-4033-af0a-36359efe7229-kube-api-access-bgx7t\") pod \"nova-metadata-0\" (UID: \"f399940f-164b-4033-af0a-36359efe7229\") " pod="openstack/nova-metadata-0" Oct 04 04:56:36 crc kubenswrapper[4992]: I1004 04:56:36.940994 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 04:56:37 crc kubenswrapper[4992]: I1004 04:56:37.354012 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 04:56:37 crc kubenswrapper[4992]: I1004 04:56:37.521150 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f399940f-164b-4033-af0a-36359efe7229","Type":"ContainerStarted","Data":"834087eefc352c07a281d5ec877be6ffce5d19c9f5734b09ab542053dedbcf30"} Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.336023 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" path="/var/lib/kubelet/pods/b3cc7753-b742-406f-abc1-b4b1c87ec984/volumes" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.436472 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.531766 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-config-data\") pod \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.531846 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-public-tls-certs\") pod \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.531949 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-internal-tls-certs\") pod \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.531972 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-combined-ca-bundle\") pod \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.532016 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-logs\") pod \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.532119 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmrzc\" (UniqueName: \"kubernetes.io/projected/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-kube-api-access-lmrzc\") pod \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\" (UID: \"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd\") " Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.532962 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-logs" (OuterVolumeSpecName: "logs") pod "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" (UID: "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.535880 4992 generic.go:334] "Generic (PLEG): container finished" podID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerID="2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b" exitCode=0 Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.535964 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.535964 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd","Type":"ContainerDied","Data":"2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b"} Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.536104 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd","Type":"ContainerDied","Data":"7d2e189d44abf91ff34f90073903348a997d506477bb36e6fcf15af390a9c0d1"} Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.536155 4992 scope.go:117] "RemoveContainer" containerID="2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.538397 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f399940f-164b-4033-af0a-36359efe7229","Type":"ContainerStarted","Data":"69c8239eab5a3c2c708bf0fbdb1c518094f7acbbcf0fde767b362ccf8e67af81"} Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.538444 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f399940f-164b-4033-af0a-36359efe7229","Type":"ContainerStarted","Data":"2a1bfcae8f661b2a1ae371c5d51a468a0ba2da73d2e479cbcde699a6296150d3"} Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.538652 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-kube-api-access-lmrzc" (OuterVolumeSpecName: "kube-api-access-lmrzc") pod "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" (UID: "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd"). InnerVolumeSpecName "kube-api-access-lmrzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.570841 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.570815423 podStartE2EDuration="2.570815423s" podCreationTimestamp="2025-10-04 04:56:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:38.558517037 +0000 UTC m=+1432.406192515" watchObservedRunningTime="2025-10-04 04:56:38.570815423 +0000 UTC m=+1432.418490891" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.587781 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" (UID: "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.590974 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-config-data" (OuterVolumeSpecName: "config-data") pod "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" (UID: "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.596950 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" (UID: "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.600757 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" (UID: "0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.613384 4992 scope.go:117] "RemoveContainer" containerID="25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.632649 4992 scope.go:117] "RemoveContainer" containerID="2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b" Oct 04 04:56:38 crc kubenswrapper[4992]: E1004 04:56:38.633211 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b\": container with ID starting with 2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b not found: ID does not exist" containerID="2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.633253 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b"} err="failed to get container status \"2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b\": rpc error: code = NotFound desc = could not find container \"2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b\": container with ID starting with 2d72808211f0154b5e027688f0797797ef5886f26d41903bfa9de7a969406e5b not found: ID does not exist" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.633283 4992 scope.go:117] "RemoveContainer" containerID="25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.633866 4992 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.633938 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.633952 4992 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-logs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:38 crc kubenswrapper[4992]: E1004 04:56:38.633951 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee\": container with ID starting with 25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee not found: ID does not exist" containerID="25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.634002 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee"} err="failed to get container status \"25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee\": rpc error: code = NotFound desc = could not find container \"25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee\": container with ID starting with 25c9254317310dbb711ed148bed199103b4d0e5c550fe6e0cff2e88d05b142ee not found: ID does not exist" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.633962 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmrzc\" (UniqueName: \"kubernetes.io/projected/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-kube-api-access-lmrzc\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.634050 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.634061 4992 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.905136 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.925030 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.940098 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:38 crc kubenswrapper[4992]: E1004 04:56:38.940642 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-log" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.940663 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-log" Oct 04 04:56:38 crc kubenswrapper[4992]: E1004 04:56:38.940681 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-api" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.940690 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-api" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.940955 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-log" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.940984 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" containerName="nova-api-api" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.945749 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.956492 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.956730 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.956859 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 04:56:38 crc kubenswrapper[4992]: I1004 04:56:38.957803 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.040715 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-public-tls-certs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.040788 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czdbz\" (UniqueName: \"kubernetes.io/projected/23d3720b-cfd2-4513-93c0-6f9256dda97a-kube-api-access-czdbz\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.040817 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.040837 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-config-data\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.040914 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d3720b-cfd2-4513-93c0-6f9256dda97a-logs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.040939 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.142082 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-public-tls-certs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.142164 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czdbz\" (UniqueName: \"kubernetes.io/projected/23d3720b-cfd2-4513-93c0-6f9256dda97a-kube-api-access-czdbz\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.142206 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.142234 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-config-data\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.142271 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d3720b-cfd2-4513-93c0-6f9256dda97a-logs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.142303 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.144300 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d3720b-cfd2-4513-93c0-6f9256dda97a-logs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.146682 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.146877 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.153315 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-config-data\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.153763 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23d3720b-cfd2-4513-93c0-6f9256dda97a-public-tls-certs\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.159608 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czdbz\" (UniqueName: \"kubernetes.io/projected/23d3720b-cfd2-4513-93c0-6f9256dda97a-kube-api-access-czdbz\") pod \"nova-api-0\" (UID: \"23d3720b-cfd2-4513-93c0-6f9256dda97a\") " pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.266848 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.733681 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 04:56:39 crc kubenswrapper[4992]: W1004 04:56:39.740101 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23d3720b_cfd2_4513_93c0_6f9256dda97a.slice/crio-e2ac4213420d37f4a2e560d694af0c46c62b069aa701929f936ae886c1e08a0d WatchSource:0}: Error finding container e2ac4213420d37f4a2e560d694af0c46c62b069aa701929f936ae886c1e08a0d: Status 404 returned error can't find the container with id e2ac4213420d37f4a2e560d694af0c46c62b069aa701929f936ae886c1e08a0d Oct 04 04:56:39 crc kubenswrapper[4992]: I1004 04:56:39.891975 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 04:56:40 crc kubenswrapper[4992]: I1004 04:56:40.330215 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd" path="/var/lib/kubelet/pods/0fca9cad-13e6-4ff2-a5c4-8c3e31133cfd/volumes" Oct 04 04:56:40 crc kubenswrapper[4992]: I1004 04:56:40.560562 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23d3720b-cfd2-4513-93c0-6f9256dda97a","Type":"ContainerStarted","Data":"b257ebd0874cba0ff574c951fcb0efedc2ae05ed97dcd16f33d7bb1bdb23fa38"} Oct 04 04:56:40 crc kubenswrapper[4992]: I1004 04:56:40.560607 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23d3720b-cfd2-4513-93c0-6f9256dda97a","Type":"ContainerStarted","Data":"42bfd65b18c2743da7a20bd7c07364ac1398eb6fce507f0fb124ab346f80aa70"} Oct 04 04:56:40 crc kubenswrapper[4992]: I1004 04:56:40.560617 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23d3720b-cfd2-4513-93c0-6f9256dda97a","Type":"ContainerStarted","Data":"e2ac4213420d37f4a2e560d694af0c46c62b069aa701929f936ae886c1e08a0d"} Oct 04 04:56:40 crc kubenswrapper[4992]: I1004 04:56:40.582749 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.582727426 podStartE2EDuration="2.582727426s" podCreationTimestamp="2025-10-04 04:56:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:56:40.579045448 +0000 UTC m=+1434.426720916" watchObservedRunningTime="2025-10-04 04:56:40.582727426 +0000 UTC m=+1434.430402894" Oct 04 04:56:41 crc kubenswrapper[4992]: I1004 04:56:41.218205 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:41 crc kubenswrapper[4992]: I1004 04:56:41.218240 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b3cc7753-b742-406f-abc1-b4b1c87ec984" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:41 crc kubenswrapper[4992]: I1004 04:56:41.941809 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:56:41 crc kubenswrapper[4992]: I1004 04:56:41.941900 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 04:56:44 crc kubenswrapper[4992]: I1004 04:56:44.892127 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 04:56:44 crc kubenswrapper[4992]: I1004 04:56:44.921755 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 04:56:45 crc kubenswrapper[4992]: I1004 04:56:45.627505 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 04:56:46 crc kubenswrapper[4992]: I1004 04:56:46.941613 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:56:46 crc kubenswrapper[4992]: I1004 04:56:46.941987 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 04:56:47 crc kubenswrapper[4992]: I1004 04:56:47.958569 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f399940f-164b-4033-af0a-36359efe7229" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:47 crc kubenswrapper[4992]: I1004 04:56:47.958584 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f399940f-164b-4033-af0a-36359efe7229" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:49 crc kubenswrapper[4992]: I1004 04:56:49.268340 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:56:49 crc kubenswrapper[4992]: I1004 04:56:49.268432 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 04:56:50 crc kubenswrapper[4992]: I1004 04:56:50.280675 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23d3720b-cfd2-4513-93c0-6f9256dda97a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:50 crc kubenswrapper[4992]: I1004 04:56:50.280678 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23d3720b-cfd2-4513-93c0-6f9256dda97a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 04:56:50 crc kubenswrapper[4992]: I1004 04:56:50.894880 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:56:50 crc kubenswrapper[4992]: I1004 04:56:50.894937 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:56:51 crc kubenswrapper[4992]: I1004 04:56:51.698815 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 04:56:56 crc kubenswrapper[4992]: I1004 04:56:56.947407 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:56:56 crc kubenswrapper[4992]: I1004 04:56:56.948024 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 04:56:56 crc kubenswrapper[4992]: I1004 04:56:56.954706 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:56:56 crc kubenswrapper[4992]: I1004 04:56:56.958226 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 04:56:59 crc kubenswrapper[4992]: I1004 04:56:59.326511 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:56:59 crc kubenswrapper[4992]: I1004 04:56:59.327415 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:56:59 crc kubenswrapper[4992]: I1004 04:56:59.329204 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 04:56:59 crc kubenswrapper[4992]: I1004 04:56:59.336246 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:56:59 crc kubenswrapper[4992]: I1004 04:56:59.718503 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 04:56:59 crc kubenswrapper[4992]: I1004 04:56:59.726017 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 04:57:08 crc kubenswrapper[4992]: I1004 04:57:08.078514 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:57:08 crc kubenswrapper[4992]: I1004 04:57:08.906157 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:57:12 crc kubenswrapper[4992]: I1004 04:57:12.205648 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerName="rabbitmq" containerID="cri-o://fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15" gracePeriod=604796 Oct 04 04:57:13 crc kubenswrapper[4992]: I1004 04:57:13.266597 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerName="rabbitmq" containerID="cri-o://328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049" gracePeriod=604796 Oct 04 04:57:15 crc kubenswrapper[4992]: I1004 04:57:15.772398 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Oct 04 04:57:16 crc kubenswrapper[4992]: I1004 04:57:16.050128 4992 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.805010 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890273 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-server-conf\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890317 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890347 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-tls\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890404 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c7fa3bc-a429-4f65-aac7-a08293b802eb-pod-info\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890508 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-plugins-conf\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890535 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c7fa3bc-a429-4f65-aac7-a08293b802eb-erlang-cookie-secret\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890608 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-confd\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890663 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-erlang-cookie\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890723 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzp7z\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-kube-api-access-nzp7z\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890768 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-config-data\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.890827 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-plugins\") pod \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\" (UID: \"0c7fa3bc-a429-4f65-aac7-a08293b802eb\") " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.891816 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.893036 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.893537 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.902756 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-kube-api-access-nzp7z" (OuterVolumeSpecName: "kube-api-access-nzp7z") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "kube-api-access-nzp7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.902770 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0c7fa3bc-a429-4f65-aac7-a08293b802eb-pod-info" (OuterVolumeSpecName: "pod-info") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.903254 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c7fa3bc-a429-4f65-aac7-a08293b802eb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.910050 4992 generic.go:334] "Generic (PLEG): container finished" podID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerID="fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15" exitCode=0 Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.910102 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0c7fa3bc-a429-4f65-aac7-a08293b802eb","Type":"ContainerDied","Data":"fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15"} Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.910134 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0c7fa3bc-a429-4f65-aac7-a08293b802eb","Type":"ContainerDied","Data":"a8ebe42e01a8453829f66bffc8b9ce8022fe9017b94dea8cbae81582316cbcd6"} Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.910153 4992 scope.go:117] "RemoveContainer" containerID="fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.910155 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.918400 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.918722 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.954861 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-config-data" (OuterVolumeSpecName: "config-data") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.989490 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-server-conf" (OuterVolumeSpecName: "server-conf") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993789 4992 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993813 4992 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c7fa3bc-a429-4f65-aac7-a08293b802eb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993826 4992 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993836 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzp7z\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-kube-api-access-nzp7z\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993845 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993852 4992 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993860 4992 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c7fa3bc-a429-4f65-aac7-a08293b802eb-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993880 4992 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993891 4992 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:18 crc kubenswrapper[4992]: I1004 04:57:18.993902 4992 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c7fa3bc-a429-4f65-aac7-a08293b802eb-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.046908 4992 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.051065 4992 scope.go:117] "RemoveContainer" containerID="f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.082863 4992 scope.go:117] "RemoveContainer" containerID="fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15" Oct 04 04:57:19 crc kubenswrapper[4992]: E1004 04:57:19.083461 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15\": container with ID starting with fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15 not found: ID does not exist" containerID="fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.083500 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15"} err="failed to get container status \"fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15\": rpc error: code = NotFound desc = could not find container \"fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15\": container with ID starting with fb5d1f98de9b7aa7af92c27619431d9c0f9953a1b9fa66d42fc1ed143e52ef15 not found: ID does not exist" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.083525 4992 scope.go:117] "RemoveContainer" containerID="f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd" Oct 04 04:57:19 crc kubenswrapper[4992]: E1004 04:57:19.083778 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd\": container with ID starting with f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd not found: ID does not exist" containerID="f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.083806 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd"} err="failed to get container status \"f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd\": rpc error: code = NotFound desc = could not find container \"f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd\": container with ID starting with f957f78e3a1c99efac1bc4efb201b18dc333ff6fdafb18f690ec93b6bef552fd not found: ID does not exist" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.095144 4992 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.101017 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0c7fa3bc-a429-4f65-aac7-a08293b802eb" (UID: "0c7fa3bc-a429-4f65-aac7-a08293b802eb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.196950 4992 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c7fa3bc-a429-4f65-aac7-a08293b802eb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.247245 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.256906 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.293112 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:57:19 crc kubenswrapper[4992]: E1004 04:57:19.293661 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerName="rabbitmq" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.293686 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerName="rabbitmq" Oct 04 04:57:19 crc kubenswrapper[4992]: E1004 04:57:19.293709 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerName="setup-container" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.293718 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerName="setup-container" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.293970 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" containerName="rabbitmq" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.295279 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.299014 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.299102 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.311490 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.312145 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.312145 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.317137 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.321225 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-h2k25" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.321250 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.331211 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-xsfxl"] Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.333017 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.342103 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.349087 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-xsfxl"] Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.401793 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.401860 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.401992 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402028 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-pod-info\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402055 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-server-conf\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402185 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402241 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402296 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2sk4\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-kube-api-access-b2sk4\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402398 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tvgs\" (UniqueName: \"kubernetes.io/projected/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-kube-api-access-4tvgs\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402434 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402472 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402503 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-config\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402551 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-svc\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402578 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402607 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402633 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-config-data\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402671 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.402748 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504163 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tvgs\" (UniqueName: \"kubernetes.io/projected/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-kube-api-access-4tvgs\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504229 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504264 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504291 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-config\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504325 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-svc\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504345 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504389 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504412 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-config-data\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504439 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504476 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504530 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504566 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504619 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504644 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-pod-info\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504673 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-server-conf\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504707 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504733 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.504764 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2sk4\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-kube-api-access-b2sk4\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.505731 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.506482 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.507132 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.507295 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.507476 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.507484 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.507524 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-config\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.508319 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-svc\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.509131 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.509216 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-config-data\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.510029 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-server-conf\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.510335 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.519490 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-pod-info\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.519496 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.519787 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.523724 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.528531 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2sk4\" (UniqueName: \"kubernetes.io/projected/764fc1cd-ff45-46fe-8a26-e2dfdb02ed71-kube-api-access-b2sk4\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.528651 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tvgs\" (UniqueName: \"kubernetes.io/projected/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-kube-api-access-4tvgs\") pod \"dnsmasq-dns-67b789f86c-xsfxl\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.547798 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71\") " pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.622419 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.657834 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.837195 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914051 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-pod-info\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914131 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-config-data\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914198 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-tls\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914280 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrrrv\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-kube-api-access-mrrrv\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914335 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-erlang-cookie-secret\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914383 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-confd\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914463 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-plugins-conf\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914486 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-plugins\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914508 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-server-conf\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914596 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.914617 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-erlang-cookie\") pod \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\" (UID: \"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52\") " Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.915407 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.915960 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.916261 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.918351 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-pod-info" (OuterVolumeSpecName: "pod-info") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.921251 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.924686 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.925175 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-kube-api-access-mrrrv" (OuterVolumeSpecName: "kube-api-access-mrrrv") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "kube-api-access-mrrrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.926399 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.931633 4992 generic.go:334] "Generic (PLEG): container finished" podID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerID="328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049" exitCode=0 Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.931713 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52","Type":"ContainerDied","Data":"328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049"} Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.931759 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c212cbf1-b9fa-40d5-b3e6-da9f74f65b52","Type":"ContainerDied","Data":"2369543569a5e75a9a5f226f2abf34b54449cb743aff22d9911cb55212b8993f"} Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.931778 4992 scope.go:117] "RemoveContainer" containerID="328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.931819 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.962177 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-config-data" (OuterVolumeSpecName: "config-data") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:19 crc kubenswrapper[4992]: I1004 04:57:19.962765 4992 scope.go:117] "RemoveContainer" containerID="c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.009080 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-server-conf" (OuterVolumeSpecName: "server-conf") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.012646 4992 scope.go:117] "RemoveContainer" containerID="328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049" Oct 04 04:57:20 crc kubenswrapper[4992]: E1004 04:57:20.013198 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049\": container with ID starting with 328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049 not found: ID does not exist" containerID="328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.013255 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049"} err="failed to get container status \"328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049\": rpc error: code = NotFound desc = could not find container \"328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049\": container with ID starting with 328e66baa56a7c4d0e0dee6d40bee2f00ed117acd6851ad0e03f9b3cdb263049 not found: ID does not exist" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.013284 4992 scope.go:117] "RemoveContainer" containerID="c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44" Oct 04 04:57:20 crc kubenswrapper[4992]: E1004 04:57:20.013736 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44\": container with ID starting with c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44 not found: ID does not exist" containerID="c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.013760 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44"} err="failed to get container status \"c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44\": rpc error: code = NotFound desc = could not find container \"c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44\": container with ID starting with c6da60bbc9c000a634f8dac23635385f2b7133d5cd23d3cd910ee2be6dfe9b44 not found: ID does not exist" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017029 4992 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017060 4992 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017073 4992 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017088 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017098 4992 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017110 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrrrv\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-kube-api-access-mrrrv\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017121 4992 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017131 4992 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017141 4992 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.017150 4992 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.052243 4992 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.066621 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" (UID: "c212cbf1-b9fa-40d5-b3e6-da9f74f65b52"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.119076 4992 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.119115 4992 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.151631 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.271724 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.280653 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:57:20 crc kubenswrapper[4992]: W1004 04:57:20.304093 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e4d774b_6d1b_41e6_9e54_1e725a7379f5.slice/crio-5de570377d41012407433d4bcfb22adbdd49e9d6df5fd25d2b4fb5501e647f0e WatchSource:0}: Error finding container 5de570377d41012407433d4bcfb22adbdd49e9d6df5fd25d2b4fb5501e647f0e: Status 404 returned error can't find the container with id 5de570377d41012407433d4bcfb22adbdd49e9d6df5fd25d2b4fb5501e647f0e Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.316832 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:57:20 crc kubenswrapper[4992]: E1004 04:57:20.326049 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerName="rabbitmq" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.326094 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerName="rabbitmq" Oct 04 04:57:20 crc kubenswrapper[4992]: E1004 04:57:20.326110 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerName="setup-container" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.326124 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerName="setup-container" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.326802 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" containerName="rabbitmq" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.329962 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.332834 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-558fj" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.334073 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.334666 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.335545 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.335738 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.335922 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.337719 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.373482 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c7fa3bc-a429-4f65-aac7-a08293b802eb" path="/var/lib/kubelet/pods/0c7fa3bc-a429-4f65-aac7-a08293b802eb/volumes" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.374590 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c212cbf1-b9fa-40d5-b3e6-da9f74f65b52" path="/var/lib/kubelet/pods/c212cbf1-b9fa-40d5-b3e6-da9f74f65b52/volumes" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.375332 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-xsfxl"] Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.375393 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.431230 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.431533 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.431682 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.431824 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.431933 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.432073 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.432167 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.432287 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps7xd\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-kube-api-access-ps7xd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.432431 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.432538 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.432674 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.534012 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.534520 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.534612 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.534712 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.534820 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.534895 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.535008 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.535092 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.535195 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps7xd\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-kube-api-access-ps7xd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.535287 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.535465 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.536022 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.537823 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.538113 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.538145 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.538546 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.539618 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.540521 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.542324 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.543023 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.545823 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.555634 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps7xd\" (UniqueName: \"kubernetes.io/projected/e3c63a70-cea7-4b0f-bf50-1f9dea7b1370-kube-api-access-ps7xd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.566948 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.657195 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.895280 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.896057 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.950097 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71","Type":"ContainerStarted","Data":"7b5d20678b3acb4f65c434aee7aa98e17a8ec5cb310d5365c532a0829ee6faba"} Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.952452 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" event={"ID":"9e4d774b-6d1b-41e6-9e54-1e725a7379f5","Type":"ContainerStarted","Data":"9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a"} Oct 04 04:57:20 crc kubenswrapper[4992]: I1004 04:57:20.952479 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" event={"ID":"9e4d774b-6d1b-41e6-9e54-1e725a7379f5","Type":"ContainerStarted","Data":"5de570377d41012407433d4bcfb22adbdd49e9d6df5fd25d2b4fb5501e647f0e"} Oct 04 04:57:21 crc kubenswrapper[4992]: I1004 04:57:21.137548 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 04:57:21 crc kubenswrapper[4992]: I1004 04:57:21.962229 4992 generic.go:334] "Generic (PLEG): container finished" podID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" containerID="9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a" exitCode=0 Oct 04 04:57:21 crc kubenswrapper[4992]: I1004 04:57:21.962304 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" event={"ID":"9e4d774b-6d1b-41e6-9e54-1e725a7379f5","Type":"ContainerDied","Data":"9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a"} Oct 04 04:57:21 crc kubenswrapper[4992]: I1004 04:57:21.965451 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71","Type":"ContainerStarted","Data":"cda596e43540881e5b56cb4c9371e7671dbb8b6eff0f0e757ca844efa6eaad38"} Oct 04 04:57:21 crc kubenswrapper[4992]: I1004 04:57:21.966866 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370","Type":"ContainerStarted","Data":"47a558150a2f1b290d75da3118e969baf54a867397ae895beacfa0dfae7b670c"} Oct 04 04:57:22 crc kubenswrapper[4992]: I1004 04:57:22.977704 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" event={"ID":"9e4d774b-6d1b-41e6-9e54-1e725a7379f5","Type":"ContainerStarted","Data":"28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a"} Oct 04 04:57:22 crc kubenswrapper[4992]: I1004 04:57:22.978147 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:22 crc kubenswrapper[4992]: I1004 04:57:22.979532 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370","Type":"ContainerStarted","Data":"b6099e1eabf9c4f79c02269a40d979e1458d6bd0e71b3a08c58965c701475aea"} Oct 04 04:57:23 crc kubenswrapper[4992]: I1004 04:57:23.000861 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" podStartSLOduration=4.000841769 podStartE2EDuration="4.000841769s" podCreationTimestamp="2025-10-04 04:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:22.994392218 +0000 UTC m=+1476.842067696" watchObservedRunningTime="2025-10-04 04:57:23.000841769 +0000 UTC m=+1476.848517237" Oct 04 04:57:29 crc kubenswrapper[4992]: I1004 04:57:29.659377 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:29 crc kubenswrapper[4992]: I1004 04:57:29.744264 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-th48l"] Oct 04 04:57:29 crc kubenswrapper[4992]: I1004 04:57:29.744641 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" podUID="30b4e051-7811-4273-a749-629ec4cff7a8" containerName="dnsmasq-dns" containerID="cri-o://78b97b081cb35ea7a4e1436df09d9b89c2ba5cada56bf249e43048a561fe3eab" gracePeriod=10 Oct 04 04:57:29 crc kubenswrapper[4992]: I1004 04:57:29.873624 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-p2jdn"] Oct 04 04:57:29 crc kubenswrapper[4992]: I1004 04:57:29.875309 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:29 crc kubenswrapper[4992]: I1004 04:57:29.890457 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-p2jdn"] Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.013652 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4nlg\" (UniqueName: \"kubernetes.io/projected/74bd057e-61c1-4538-be59-e36028dca0d9-kube-api-access-z4nlg\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.013993 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-config\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.014028 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.014092 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.014111 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.014373 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.014543 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.052991 4992 generic.go:334] "Generic (PLEG): container finished" podID="30b4e051-7811-4273-a749-629ec4cff7a8" containerID="78b97b081cb35ea7a4e1436df09d9b89c2ba5cada56bf249e43048a561fe3eab" exitCode=0 Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.053039 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" event={"ID":"30b4e051-7811-4273-a749-629ec4cff7a8","Type":"ContainerDied","Data":"78b97b081cb35ea7a4e1436df09d9b89c2ba5cada56bf249e43048a561fe3eab"} Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.130278 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.130380 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.130449 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4nlg\" (UniqueName: \"kubernetes.io/projected/74bd057e-61c1-4538-be59-e36028dca0d9-kube-api-access-z4nlg\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.130495 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-config\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.130520 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.130589 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.130611 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.131666 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.132289 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.133037 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.133206 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-config\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.133679 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.134046 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74bd057e-61c1-4538-be59-e36028dca0d9-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.158035 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4nlg\" (UniqueName: \"kubernetes.io/projected/74bd057e-61c1-4538-be59-e36028dca0d9-kube-api-access-z4nlg\") pod \"dnsmasq-dns-cb6ffcf87-p2jdn\" (UID: \"74bd057e-61c1-4538-be59-e36028dca0d9\") " pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.233774 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.362000 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.540195 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drr24\" (UniqueName: \"kubernetes.io/projected/30b4e051-7811-4273-a749-629ec4cff7a8-kube-api-access-drr24\") pod \"30b4e051-7811-4273-a749-629ec4cff7a8\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.540310 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-svc\") pod \"30b4e051-7811-4273-a749-629ec4cff7a8\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.540350 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-swift-storage-0\") pod \"30b4e051-7811-4273-a749-629ec4cff7a8\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.540393 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-sb\") pod \"30b4e051-7811-4273-a749-629ec4cff7a8\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.540485 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-nb\") pod \"30b4e051-7811-4273-a749-629ec4cff7a8\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.540525 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-config\") pod \"30b4e051-7811-4273-a749-629ec4cff7a8\" (UID: \"30b4e051-7811-4273-a749-629ec4cff7a8\") " Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.578633 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30b4e051-7811-4273-a749-629ec4cff7a8-kube-api-access-drr24" (OuterVolumeSpecName: "kube-api-access-drr24") pod "30b4e051-7811-4273-a749-629ec4cff7a8" (UID: "30b4e051-7811-4273-a749-629ec4cff7a8"). InnerVolumeSpecName "kube-api-access-drr24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.644075 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drr24\" (UniqueName: \"kubernetes.io/projected/30b4e051-7811-4273-a749-629ec4cff7a8-kube-api-access-drr24\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.679702 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-config" (OuterVolumeSpecName: "config") pod "30b4e051-7811-4273-a749-629ec4cff7a8" (UID: "30b4e051-7811-4273-a749-629ec4cff7a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.713962 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "30b4e051-7811-4273-a749-629ec4cff7a8" (UID: "30b4e051-7811-4273-a749-629ec4cff7a8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.720069 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "30b4e051-7811-4273-a749-629ec4cff7a8" (UID: "30b4e051-7811-4273-a749-629ec4cff7a8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.724098 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "30b4e051-7811-4273-a749-629ec4cff7a8" (UID: "30b4e051-7811-4273-a749-629ec4cff7a8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.726667 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "30b4e051-7811-4273-a749-629ec4cff7a8" (UID: "30b4e051-7811-4273-a749-629ec4cff7a8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.737996 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-p2jdn"] Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.745825 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.745846 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.745857 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.745867 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:30 crc kubenswrapper[4992]: I1004 04:57:30.745876 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30b4e051-7811-4273-a749-629ec4cff7a8-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.061203 4992 generic.go:334] "Generic (PLEG): container finished" podID="74bd057e-61c1-4538-be59-e36028dca0d9" containerID="f44cea6f2c41449e1864c1858fd60ed5b5b50cae167d2d9fb279bbb85b7cba1c" exitCode=0 Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.061274 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" event={"ID":"74bd057e-61c1-4538-be59-e36028dca0d9","Type":"ContainerDied","Data":"f44cea6f2c41449e1864c1858fd60ed5b5b50cae167d2d9fb279bbb85b7cba1c"} Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.061302 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" event={"ID":"74bd057e-61c1-4538-be59-e36028dca0d9","Type":"ContainerStarted","Data":"d5c210f461b3c20c419dc0d189012cc71cb8ce6b4b694209bd8ce35fa2ac24db"} Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.064261 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" event={"ID":"30b4e051-7811-4273-a749-629ec4cff7a8","Type":"ContainerDied","Data":"9d134cdeeb17ebb0b7b672a4a2c5ad347afc47e9902d86a5da10a320e9b78a28"} Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.064303 4992 scope.go:117] "RemoveContainer" containerID="78b97b081cb35ea7a4e1436df09d9b89c2ba5cada56bf249e43048a561fe3eab" Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.064348 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-th48l" Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.241769 4992 scope.go:117] "RemoveContainer" containerID="dc9119f28a854c2cf653ff0929786845541ee6e0bad370943f4adb60e304a67d" Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.276847 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-th48l"] Oct 04 04:57:31 crc kubenswrapper[4992]: I1004 04:57:31.285310 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-th48l"] Oct 04 04:57:32 crc kubenswrapper[4992]: I1004 04:57:32.073331 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" event={"ID":"74bd057e-61c1-4538-be59-e36028dca0d9","Type":"ContainerStarted","Data":"b3b108846bfb7f18c719828af76bdc54c105bf164d101efd8aed01a35d99d203"} Oct 04 04:57:32 crc kubenswrapper[4992]: I1004 04:57:32.073496 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:32 crc kubenswrapper[4992]: I1004 04:57:32.112987 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" podStartSLOduration=3.112958525 podStartE2EDuration="3.112958525s" podCreationTimestamp="2025-10-04 04:57:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:32.090469749 +0000 UTC m=+1485.938145217" watchObservedRunningTime="2025-10-04 04:57:32.112958525 +0000 UTC m=+1485.960634003" Oct 04 04:57:32 crc kubenswrapper[4992]: I1004 04:57:32.332100 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b4e051-7811-4273-a749-629ec4cff7a8" path="/var/lib/kubelet/pods/30b4e051-7811-4273-a749-629ec4cff7a8/volumes" Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.235568 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-p2jdn" Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.349640 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-xsfxl"] Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.350105 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" podUID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" containerName="dnsmasq-dns" containerID="cri-o://28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a" gracePeriod=10 Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.840149 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.942918 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-config\") pod \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.943050 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-openstack-edpm-ipam\") pod \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.943091 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-sb\") pod \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.943211 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-nb\") pod \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.943270 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-swift-storage-0\") pod \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.943322 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-svc\") pod \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.943381 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tvgs\" (UniqueName: \"kubernetes.io/projected/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-kube-api-access-4tvgs\") pod \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\" (UID: \"9e4d774b-6d1b-41e6-9e54-1e725a7379f5\") " Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.950144 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-kube-api-access-4tvgs" (OuterVolumeSpecName: "kube-api-access-4tvgs") pod "9e4d774b-6d1b-41e6-9e54-1e725a7379f5" (UID: "9e4d774b-6d1b-41e6-9e54-1e725a7379f5"). InnerVolumeSpecName "kube-api-access-4tvgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.995945 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e4d774b-6d1b-41e6-9e54-1e725a7379f5" (UID: "9e4d774b-6d1b-41e6-9e54-1e725a7379f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.996931 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "9e4d774b-6d1b-41e6-9e54-1e725a7379f5" (UID: "9e4d774b-6d1b-41e6-9e54-1e725a7379f5"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:40 crc kubenswrapper[4992]: I1004 04:57:40.999868 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e4d774b-6d1b-41e6-9e54-1e725a7379f5" (UID: "9e4d774b-6d1b-41e6-9e54-1e725a7379f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.012348 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e4d774b-6d1b-41e6-9e54-1e725a7379f5" (UID: "9e4d774b-6d1b-41e6-9e54-1e725a7379f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.013114 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e4d774b-6d1b-41e6-9e54-1e725a7379f5" (UID: "9e4d774b-6d1b-41e6-9e54-1e725a7379f5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.022866 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-config" (OuterVolumeSpecName: "config") pod "9e4d774b-6d1b-41e6-9e54-1e725a7379f5" (UID: "9e4d774b-6d1b-41e6-9e54-1e725a7379f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.046259 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.046324 4992 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.046335 4992 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.046348 4992 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.046372 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tvgs\" (UniqueName: \"kubernetes.io/projected/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-kube-api-access-4tvgs\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.046383 4992 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.046392 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9e4d774b-6d1b-41e6-9e54-1e725a7379f5-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.171443 4992 generic.go:334] "Generic (PLEG): container finished" podID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" containerID="28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a" exitCode=0 Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.171486 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" event={"ID":"9e4d774b-6d1b-41e6-9e54-1e725a7379f5","Type":"ContainerDied","Data":"28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a"} Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.171518 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" event={"ID":"9e4d774b-6d1b-41e6-9e54-1e725a7379f5","Type":"ContainerDied","Data":"5de570377d41012407433d4bcfb22adbdd49e9d6df5fd25d2b4fb5501e647f0e"} Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.171535 4992 scope.go:117] "RemoveContainer" containerID="28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.171528 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-xsfxl" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.194106 4992 scope.go:117] "RemoveContainer" containerID="9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.211510 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-xsfxl"] Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.222281 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-xsfxl"] Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.232290 4992 scope.go:117] "RemoveContainer" containerID="28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a" Oct 04 04:57:41 crc kubenswrapper[4992]: E1004 04:57:41.232982 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a\": container with ID starting with 28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a not found: ID does not exist" containerID="28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.233016 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a"} err="failed to get container status \"28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a\": rpc error: code = NotFound desc = could not find container \"28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a\": container with ID starting with 28b0ccc381bb451ce75c78c32c94908ebd792745cb1f15c26920df3127106d5a not found: ID does not exist" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.233042 4992 scope.go:117] "RemoveContainer" containerID="9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a" Oct 04 04:57:41 crc kubenswrapper[4992]: E1004 04:57:41.233763 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a\": container with ID starting with 9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a not found: ID does not exist" containerID="9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a" Oct 04 04:57:41 crc kubenswrapper[4992]: I1004 04:57:41.233794 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a"} err="failed to get container status \"9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a\": rpc error: code = NotFound desc = could not find container \"9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a\": container with ID starting with 9703dd6b4dd993b7787be62eff93db5cfe849a440a1826cd45198ec7e15d5b8a not found: ID does not exist" Oct 04 04:57:42 crc kubenswrapper[4992]: I1004 04:57:42.330799 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" path="/var/lib/kubelet/pods/9e4d774b-6d1b-41e6-9e54-1e725a7379f5/volumes" Oct 04 04:57:50 crc kubenswrapper[4992]: I1004 04:57:50.893997 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 04:57:50 crc kubenswrapper[4992]: I1004 04:57:50.894756 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 04:57:50 crc kubenswrapper[4992]: I1004 04:57:50.894812 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 04:57:50 crc kubenswrapper[4992]: I1004 04:57:50.895600 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b7858b04a08dbf3c2f36756b5a0b3054268a106fb6be0797a45494ff8fb3a8ad"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 04:57:50 crc kubenswrapper[4992]: I1004 04:57:50.895666 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://b7858b04a08dbf3c2f36756b5a0b3054268a106fb6be0797a45494ff8fb3a8ad" gracePeriod=600 Oct 04 04:57:51 crc kubenswrapper[4992]: I1004 04:57:51.268014 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="b7858b04a08dbf3c2f36756b5a0b3054268a106fb6be0797a45494ff8fb3a8ad" exitCode=0 Oct 04 04:57:51 crc kubenswrapper[4992]: I1004 04:57:51.268080 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"b7858b04a08dbf3c2f36756b5a0b3054268a106fb6be0797a45494ff8fb3a8ad"} Oct 04 04:57:51 crc kubenswrapper[4992]: I1004 04:57:51.268315 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817"} Oct 04 04:57:51 crc kubenswrapper[4992]: I1004 04:57:51.268332 4992 scope.go:117] "RemoveContainer" containerID="5c66b07d0127863fb32b4550ae384badc1dd16f486eaff1e71f5b9fa6a2b2449" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.718662 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9"] Oct 04 04:57:53 crc kubenswrapper[4992]: E1004 04:57:53.719485 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b4e051-7811-4273-a749-629ec4cff7a8" containerName="init" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.719501 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b4e051-7811-4273-a749-629ec4cff7a8" containerName="init" Oct 04 04:57:53 crc kubenswrapper[4992]: E1004 04:57:53.719517 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" containerName="dnsmasq-dns" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.719526 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" containerName="dnsmasq-dns" Oct 04 04:57:53 crc kubenswrapper[4992]: E1004 04:57:53.719569 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" containerName="init" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.719577 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" containerName="init" Oct 04 04:57:53 crc kubenswrapper[4992]: E1004 04:57:53.719591 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30b4e051-7811-4273-a749-629ec4cff7a8" containerName="dnsmasq-dns" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.719600 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="30b4e051-7811-4273-a749-629ec4cff7a8" containerName="dnsmasq-dns" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.719838 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="30b4e051-7811-4273-a749-629ec4cff7a8" containerName="dnsmasq-dns" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.719867 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e4d774b-6d1b-41e6-9e54-1e725a7379f5" containerName="dnsmasq-dns" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.720693 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.722925 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.723192 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.723202 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.726014 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.737951 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9"] Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.887603 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.887957 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqnjv\" (UniqueName: \"kubernetes.io/projected/e910a377-6253-4eda-8850-6e1f2f25f99f-kube-api-access-nqnjv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.887991 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.888126 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.989812 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqnjv\" (UniqueName: \"kubernetes.io/projected/e910a377-6253-4eda-8850-6e1f2f25f99f-kube-api-access-nqnjv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.990176 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.990451 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.990631 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.997761 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.997821 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:53 crc kubenswrapper[4992]: I1004 04:57:53.998474 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:54 crc kubenswrapper[4992]: I1004 04:57:54.009568 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqnjv\" (UniqueName: \"kubernetes.io/projected/e910a377-6253-4eda-8850-6e1f2f25f99f-kube-api-access-nqnjv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:54 crc kubenswrapper[4992]: I1004 04:57:54.049005 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:57:54 crc kubenswrapper[4992]: I1004 04:57:54.305886 4992 generic.go:334] "Generic (PLEG): container finished" podID="764fc1cd-ff45-46fe-8a26-e2dfdb02ed71" containerID="cda596e43540881e5b56cb4c9371e7671dbb8b6eff0f0e757ca844efa6eaad38" exitCode=0 Oct 04 04:57:54 crc kubenswrapper[4992]: I1004 04:57:54.306503 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71","Type":"ContainerDied","Data":"cda596e43540881e5b56cb4c9371e7671dbb8b6eff0f0e757ca844efa6eaad38"} Oct 04 04:57:55 crc kubenswrapper[4992]: I1004 04:57:54.627102 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9"] Oct 04 04:57:55 crc kubenswrapper[4992]: W1004 04:57:54.635266 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode910a377_6253_4eda_8850_6e1f2f25f99f.slice/crio-673fcca00cb1b47d4688b177bb50418cfbaa37a8970067c21921f18c1941a527 WatchSource:0}: Error finding container 673fcca00cb1b47d4688b177bb50418cfbaa37a8970067c21921f18c1941a527: Status 404 returned error can't find the container with id 673fcca00cb1b47d4688b177bb50418cfbaa37a8970067c21921f18c1941a527 Oct 04 04:57:55 crc kubenswrapper[4992]: I1004 04:57:55.316623 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" event={"ID":"e910a377-6253-4eda-8850-6e1f2f25f99f","Type":"ContainerStarted","Data":"673fcca00cb1b47d4688b177bb50418cfbaa37a8970067c21921f18c1941a527"} Oct 04 04:57:55 crc kubenswrapper[4992]: I1004 04:57:55.321395 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"764fc1cd-ff45-46fe-8a26-e2dfdb02ed71","Type":"ContainerStarted","Data":"5f33888a843639fefd09d39648235a4f242f675c203aae3b5e32f491a69c1893"} Oct 04 04:57:55 crc kubenswrapper[4992]: I1004 04:57:55.321613 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 04:57:55 crc kubenswrapper[4992]: I1004 04:57:55.324840 4992 generic.go:334] "Generic (PLEG): container finished" podID="e3c63a70-cea7-4b0f-bf50-1f9dea7b1370" containerID="b6099e1eabf9c4f79c02269a40d979e1458d6bd0e71b3a08c58965c701475aea" exitCode=0 Oct 04 04:57:55 crc kubenswrapper[4992]: I1004 04:57:55.324873 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370","Type":"ContainerDied","Data":"b6099e1eabf9c4f79c02269a40d979e1458d6bd0e71b3a08c58965c701475aea"} Oct 04 04:57:55 crc kubenswrapper[4992]: I1004 04:57:55.349180 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.349160638 podStartE2EDuration="36.349160638s" podCreationTimestamp="2025-10-04 04:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:55.349037345 +0000 UTC m=+1509.196712803" watchObservedRunningTime="2025-10-04 04:57:55.349160638 +0000 UTC m=+1509.196836106" Oct 04 04:57:56 crc kubenswrapper[4992]: I1004 04:57:56.338425 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e3c63a70-cea7-4b0f-bf50-1f9dea7b1370","Type":"ContainerStarted","Data":"b769d79df8067357bf8b0ef9753e3727cbf45a9d9e819297206bcc9bca438546"} Oct 04 04:57:56 crc kubenswrapper[4992]: I1004 04:57:56.338995 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:57:56 crc kubenswrapper[4992]: I1004 04:57:56.363611 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.363591589 podStartE2EDuration="36.363591589s" podCreationTimestamp="2025-10-04 04:57:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 04:57:56.360125097 +0000 UTC m=+1510.207800575" watchObservedRunningTime="2025-10-04 04:57:56.363591589 +0000 UTC m=+1510.211267077" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.396710 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kbtds"] Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.399063 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.412668 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kbtds"] Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.467851 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h6pr\" (UniqueName: \"kubernetes.io/projected/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-kube-api-access-4h6pr\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.468169 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-catalog-content\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.468218 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-utilities\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.569485 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h6pr\" (UniqueName: \"kubernetes.io/projected/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-kube-api-access-4h6pr\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.569641 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-catalog-content\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.569683 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-utilities\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.570199 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-catalog-content\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.570281 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-utilities\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.599821 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h6pr\" (UniqueName: \"kubernetes.io/projected/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-kube-api-access-4h6pr\") pod \"redhat-operators-kbtds\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:57 crc kubenswrapper[4992]: I1004 04:57:57.718045 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:57:58 crc kubenswrapper[4992]: I1004 04:57:58.272753 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kbtds"] Oct 04 04:58:04 crc kubenswrapper[4992]: I1004 04:58:04.418315 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbtds" event={"ID":"1bc73be3-9dfc-4e74-8fa8-51522175c7bb","Type":"ContainerStarted","Data":"29ad5e8bd0c0451e87e143927954f3ad67b7c0b0f95df8d6e8ee7ac15c34d5ec"} Oct 04 04:58:07 crc kubenswrapper[4992]: E1004 04:58:07.716239 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Oct 04 04:58:07 crc kubenswrapper[4992]: E1004 04:58:07.716935 4992 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 04 04:58:07 crc kubenswrapper[4992]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Oct 04 04:58:07 crc kubenswrapper[4992]: - hosts: all Oct 04 04:58:07 crc kubenswrapper[4992]: strategy: linear Oct 04 04:58:07 crc kubenswrapper[4992]: tasks: Oct 04 04:58:07 crc kubenswrapper[4992]: - name: Enable podified-repos Oct 04 04:58:07 crc kubenswrapper[4992]: become: true Oct 04 04:58:07 crc kubenswrapper[4992]: ansible.builtin.shell: | Oct 04 04:58:07 crc kubenswrapper[4992]: set -euxo pipefail Oct 04 04:58:07 crc kubenswrapper[4992]: pushd /var/tmp Oct 04 04:58:07 crc kubenswrapper[4992]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Oct 04 04:58:07 crc kubenswrapper[4992]: pushd repo-setup-main Oct 04 04:58:07 crc kubenswrapper[4992]: python3 -m venv ./venv Oct 04 04:58:07 crc kubenswrapper[4992]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Oct 04 04:58:07 crc kubenswrapper[4992]: ./venv/bin/repo-setup current-podified -b antelope Oct 04 04:58:07 crc kubenswrapper[4992]: popd Oct 04 04:58:07 crc kubenswrapper[4992]: rm -rf repo-setup-main Oct 04 04:58:07 crc kubenswrapper[4992]: Oct 04 04:58:07 crc kubenswrapper[4992]: Oct 04 04:58:07 crc kubenswrapper[4992]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Oct 04 04:58:07 crc kubenswrapper[4992]: edpm_override_hosts: openstack-edpm-ipam Oct 04 04:58:07 crc kubenswrapper[4992]: edpm_service_type: repo-setup Oct 04 04:58:07 crc kubenswrapper[4992]: Oct 04 04:58:07 crc kubenswrapper[4992]: Oct 04 04:58:07 crc kubenswrapper[4992]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nqnjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9_openstack(e910a377-6253-4eda-8850-6e1f2f25f99f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Oct 04 04:58:07 crc kubenswrapper[4992]: > logger="UnhandledError" Oct 04 04:58:07 crc kubenswrapper[4992]: E1004 04:58:07.718581 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" podUID="e910a377-6253-4eda-8850-6e1f2f25f99f" Oct 04 04:58:08 crc kubenswrapper[4992]: I1004 04:58:08.454741 4992 generic.go:334] "Generic (PLEG): container finished" podID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerID="0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e" exitCode=0 Oct 04 04:58:08 crc kubenswrapper[4992]: I1004 04:58:08.454796 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbtds" event={"ID":"1bc73be3-9dfc-4e74-8fa8-51522175c7bb","Type":"ContainerDied","Data":"0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e"} Oct 04 04:58:08 crc kubenswrapper[4992]: E1004 04:58:08.456780 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" podUID="e910a377-6253-4eda-8850-6e1f2f25f99f" Oct 04 04:58:09 crc kubenswrapper[4992]: I1004 04:58:09.626590 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 04:58:10 crc kubenswrapper[4992]: I1004 04:58:10.472941 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbtds" event={"ID":"1bc73be3-9dfc-4e74-8fa8-51522175c7bb","Type":"ContainerStarted","Data":"779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b"} Oct 04 04:58:10 crc kubenswrapper[4992]: I1004 04:58:10.662607 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 04:58:11 crc kubenswrapper[4992]: I1004 04:58:11.484228 4992 generic.go:334] "Generic (PLEG): container finished" podID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerID="779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b" exitCode=0 Oct 04 04:58:11 crc kubenswrapper[4992]: I1004 04:58:11.484326 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbtds" event={"ID":"1bc73be3-9dfc-4e74-8fa8-51522175c7bb","Type":"ContainerDied","Data":"779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b"} Oct 04 04:58:13 crc kubenswrapper[4992]: I1004 04:58:13.513293 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbtds" event={"ID":"1bc73be3-9dfc-4e74-8fa8-51522175c7bb","Type":"ContainerStarted","Data":"72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59"} Oct 04 04:58:13 crc kubenswrapper[4992]: I1004 04:58:13.545859 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kbtds" podStartSLOduration=12.442340304 podStartE2EDuration="16.545833721s" podCreationTimestamp="2025-10-04 04:57:57 +0000 UTC" firstStartedPulling="2025-10-04 04:58:08.456726867 +0000 UTC m=+1522.304402345" lastFinishedPulling="2025-10-04 04:58:12.560220294 +0000 UTC m=+1526.407895762" observedRunningTime="2025-10-04 04:58:13.533952736 +0000 UTC m=+1527.381628214" watchObservedRunningTime="2025-10-04 04:58:13.545833721 +0000 UTC m=+1527.393509189" Oct 04 04:58:17 crc kubenswrapper[4992]: I1004 04:58:17.718887 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:58:17 crc kubenswrapper[4992]: I1004 04:58:17.719197 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:58:17 crc kubenswrapper[4992]: I1004 04:58:17.761919 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:58:18 crc kubenswrapper[4992]: I1004 04:58:18.611369 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:58:18 crc kubenswrapper[4992]: I1004 04:58:18.665560 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kbtds"] Oct 04 04:58:20 crc kubenswrapper[4992]: I1004 04:58:20.584575 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kbtds" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerName="registry-server" containerID="cri-o://72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59" gracePeriod=2 Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.035765 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.223500 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-utilities\") pod \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.223569 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-utilities" (OuterVolumeSpecName: "utilities") pod "1bc73be3-9dfc-4e74-8fa8-51522175c7bb" (UID: "1bc73be3-9dfc-4e74-8fa8-51522175c7bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.223930 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h6pr\" (UniqueName: \"kubernetes.io/projected/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-kube-api-access-4h6pr\") pod \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.223996 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-catalog-content\") pod \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\" (UID: \"1bc73be3-9dfc-4e74-8fa8-51522175c7bb\") " Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.224852 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.229797 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-kube-api-access-4h6pr" (OuterVolumeSpecName: "kube-api-access-4h6pr") pod "1bc73be3-9dfc-4e74-8fa8-51522175c7bb" (UID: "1bc73be3-9dfc-4e74-8fa8-51522175c7bb"). InnerVolumeSpecName "kube-api-access-4h6pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.315935 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bc73be3-9dfc-4e74-8fa8-51522175c7bb" (UID: "1bc73be3-9dfc-4e74-8fa8-51522175c7bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.326647 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h6pr\" (UniqueName: \"kubernetes.io/projected/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-kube-api-access-4h6pr\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.326693 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc73be3-9dfc-4e74-8fa8-51522175c7bb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.596808 4992 generic.go:334] "Generic (PLEG): container finished" podID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerID="72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59" exitCode=0 Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.596904 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kbtds" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.596905 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbtds" event={"ID":"1bc73be3-9dfc-4e74-8fa8-51522175c7bb","Type":"ContainerDied","Data":"72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59"} Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.597277 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kbtds" event={"ID":"1bc73be3-9dfc-4e74-8fa8-51522175c7bb","Type":"ContainerDied","Data":"29ad5e8bd0c0451e87e143927954f3ad67b7c0b0f95df8d6e8ee7ac15c34d5ec"} Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.597301 4992 scope.go:117] "RemoveContainer" containerID="72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.628850 4992 scope.go:117] "RemoveContainer" containerID="779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.635744 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kbtds"] Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.645232 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kbtds"] Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.659942 4992 scope.go:117] "RemoveContainer" containerID="0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.706162 4992 scope.go:117] "RemoveContainer" containerID="72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59" Oct 04 04:58:21 crc kubenswrapper[4992]: E1004 04:58:21.706849 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59\": container with ID starting with 72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59 not found: ID does not exist" containerID="72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.706880 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59"} err="failed to get container status \"72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59\": rpc error: code = NotFound desc = could not find container \"72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59\": container with ID starting with 72468830d286230eb2dd50315048ebe633e0819103ef613cff08b594f7093f59 not found: ID does not exist" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.706906 4992 scope.go:117] "RemoveContainer" containerID="779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b" Oct 04 04:58:21 crc kubenswrapper[4992]: E1004 04:58:21.707161 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b\": container with ID starting with 779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b not found: ID does not exist" containerID="779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.707183 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b"} err="failed to get container status \"779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b\": rpc error: code = NotFound desc = could not find container \"779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b\": container with ID starting with 779a26c35ad233f2062a869cb17cb1356ec09de29ab4ed55cf9324eab251669b not found: ID does not exist" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.707195 4992 scope.go:117] "RemoveContainer" containerID="0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e" Oct 04 04:58:21 crc kubenswrapper[4992]: E1004 04:58:21.707679 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e\": container with ID starting with 0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e not found: ID does not exist" containerID="0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e" Oct 04 04:58:21 crc kubenswrapper[4992]: I1004 04:58:21.707775 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e"} err="failed to get container status \"0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e\": rpc error: code = NotFound desc = could not find container \"0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e\": container with ID starting with 0aa97ae0abfbe7380c68e3d9d992a0d3e89aca42e608b02068141b4ef224744e not found: ID does not exist" Oct 04 04:58:22 crc kubenswrapper[4992]: I1004 04:58:22.331141 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" path="/var/lib/kubelet/pods/1bc73be3-9dfc-4e74-8fa8-51522175c7bb/volumes" Oct 04 04:58:24 crc kubenswrapper[4992]: I1004 04:58:24.221710 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:58:24 crc kubenswrapper[4992]: I1004 04:58:24.628109 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" event={"ID":"e910a377-6253-4eda-8850-6e1f2f25f99f","Type":"ContainerStarted","Data":"6519323852d69a59e5fefa0fdc24f1b35707b577cda302fb8e3ed14da2ed5230"} Oct 04 04:58:24 crc kubenswrapper[4992]: I1004 04:58:24.652612 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" podStartSLOduration=2.071481958 podStartE2EDuration="31.652588852s" podCreationTimestamp="2025-10-04 04:57:53 +0000 UTC" firstStartedPulling="2025-10-04 04:57:54.637855593 +0000 UTC m=+1508.485531051" lastFinishedPulling="2025-10-04 04:58:24.218962477 +0000 UTC m=+1538.066637945" observedRunningTime="2025-10-04 04:58:24.642112214 +0000 UTC m=+1538.489787682" watchObservedRunningTime="2025-10-04 04:58:24.652588852 +0000 UTC m=+1538.500264320" Oct 04 04:58:36 crc kubenswrapper[4992]: I1004 04:58:36.755924 4992 generic.go:334] "Generic (PLEG): container finished" podID="e910a377-6253-4eda-8850-6e1f2f25f99f" containerID="6519323852d69a59e5fefa0fdc24f1b35707b577cda302fb8e3ed14da2ed5230" exitCode=0 Oct 04 04:58:36 crc kubenswrapper[4992]: I1004 04:58:36.756004 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" event={"ID":"e910a377-6253-4eda-8850-6e1f2f25f99f","Type":"ContainerDied","Data":"6519323852d69a59e5fefa0fdc24f1b35707b577cda302fb8e3ed14da2ed5230"} Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.190574 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.363675 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-repo-setup-combined-ca-bundle\") pod \"e910a377-6253-4eda-8850-6e1f2f25f99f\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.363898 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-ssh-key\") pod \"e910a377-6253-4eda-8850-6e1f2f25f99f\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.363978 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-inventory\") pod \"e910a377-6253-4eda-8850-6e1f2f25f99f\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.364048 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqnjv\" (UniqueName: \"kubernetes.io/projected/e910a377-6253-4eda-8850-6e1f2f25f99f-kube-api-access-nqnjv\") pod \"e910a377-6253-4eda-8850-6e1f2f25f99f\" (UID: \"e910a377-6253-4eda-8850-6e1f2f25f99f\") " Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.369796 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e910a377-6253-4eda-8850-6e1f2f25f99f" (UID: "e910a377-6253-4eda-8850-6e1f2f25f99f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.370475 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e910a377-6253-4eda-8850-6e1f2f25f99f-kube-api-access-nqnjv" (OuterVolumeSpecName: "kube-api-access-nqnjv") pod "e910a377-6253-4eda-8850-6e1f2f25f99f" (UID: "e910a377-6253-4eda-8850-6e1f2f25f99f"). InnerVolumeSpecName "kube-api-access-nqnjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.395653 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-inventory" (OuterVolumeSpecName: "inventory") pod "e910a377-6253-4eda-8850-6e1f2f25f99f" (UID: "e910a377-6253-4eda-8850-6e1f2f25f99f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.395715 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e910a377-6253-4eda-8850-6e1f2f25f99f" (UID: "e910a377-6253-4eda-8850-6e1f2f25f99f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.467953 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.468139 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.468199 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqnjv\" (UniqueName: \"kubernetes.io/projected/e910a377-6253-4eda-8850-6e1f2f25f99f-kube-api-access-nqnjv\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.468258 4992 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e910a377-6253-4eda-8850-6e1f2f25f99f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.781081 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" event={"ID":"e910a377-6253-4eda-8850-6e1f2f25f99f","Type":"ContainerDied","Data":"673fcca00cb1b47d4688b177bb50418cfbaa37a8970067c21921f18c1941a527"} Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.781139 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="673fcca00cb1b47d4688b177bb50418cfbaa37a8970067c21921f18c1941a527" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.781097 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.862742 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r"] Oct 04 04:58:38 crc kubenswrapper[4992]: E1004 04:58:38.863194 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerName="extract-content" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.863214 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerName="extract-content" Oct 04 04:58:38 crc kubenswrapper[4992]: E1004 04:58:38.863233 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e910a377-6253-4eda-8850-6e1f2f25f99f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.863243 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="e910a377-6253-4eda-8850-6e1f2f25f99f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 04:58:38 crc kubenswrapper[4992]: E1004 04:58:38.863281 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerName="extract-utilities" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.863290 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerName="extract-utilities" Oct 04 04:58:38 crc kubenswrapper[4992]: E1004 04:58:38.863313 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerName="registry-server" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.863321 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerName="registry-server" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.863602 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc73be3-9dfc-4e74-8fa8-51522175c7bb" containerName="registry-server" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.863628 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="e910a377-6253-4eda-8850-6e1f2f25f99f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.864467 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.867904 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.867980 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.867987 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.869739 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.875897 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r"] Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.976540 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.976630 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfzxw\" (UniqueName: \"kubernetes.io/projected/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-kube-api-access-sfzxw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:38 crc kubenswrapper[4992]: I1004 04:58:38.976876 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.078989 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.079140 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfzxw\" (UniqueName: \"kubernetes.io/projected/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-kube-api-access-sfzxw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.079321 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.083611 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.089813 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.099259 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfzxw\" (UniqueName: \"kubernetes.io/projected/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-kube-api-access-sfzxw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-pmt2r\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.183760 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.723533 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r"] Oct 04 04:58:39 crc kubenswrapper[4992]: I1004 04:58:39.792060 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" event={"ID":"714c50bd-a6fa-44fb-ae36-f334c9bd9d68","Type":"ContainerStarted","Data":"f6c189c77300241a55c00e07dd58b94f511c041bd39b40da6ea64bccaa5cb8b8"} Oct 04 04:58:40 crc kubenswrapper[4992]: I1004 04:58:40.806826 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" event={"ID":"714c50bd-a6fa-44fb-ae36-f334c9bd9d68","Type":"ContainerStarted","Data":"cc4b569d73cf10ea832a0b89f74f8063189c6b2da40f7ba583783414ad9dc9f5"} Oct 04 04:58:40 crc kubenswrapper[4992]: I1004 04:58:40.826259 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" podStartSLOduration=2.442869879 podStartE2EDuration="2.826240983s" podCreationTimestamp="2025-10-04 04:58:38 +0000 UTC" firstStartedPulling="2025-10-04 04:58:39.735808527 +0000 UTC m=+1553.583483995" lastFinishedPulling="2025-10-04 04:58:40.119179641 +0000 UTC m=+1553.966855099" observedRunningTime="2025-10-04 04:58:40.823216853 +0000 UTC m=+1554.670892321" watchObservedRunningTime="2025-10-04 04:58:40.826240983 +0000 UTC m=+1554.673916451" Oct 04 04:58:43 crc kubenswrapper[4992]: I1004 04:58:43.839626 4992 generic.go:334] "Generic (PLEG): container finished" podID="714c50bd-a6fa-44fb-ae36-f334c9bd9d68" containerID="cc4b569d73cf10ea832a0b89f74f8063189c6b2da40f7ba583783414ad9dc9f5" exitCode=0 Oct 04 04:58:43 crc kubenswrapper[4992]: I1004 04:58:43.839725 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" event={"ID":"714c50bd-a6fa-44fb-ae36-f334c9bd9d68","Type":"ContainerDied","Data":"cc4b569d73cf10ea832a0b89f74f8063189c6b2da40f7ba583783414ad9dc9f5"} Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.251394 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.305147 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-inventory\") pod \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.305239 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfzxw\" (UniqueName: \"kubernetes.io/projected/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-kube-api-access-sfzxw\") pod \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.305270 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-ssh-key\") pod \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\" (UID: \"714c50bd-a6fa-44fb-ae36-f334c9bd9d68\") " Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.311307 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-kube-api-access-sfzxw" (OuterVolumeSpecName: "kube-api-access-sfzxw") pod "714c50bd-a6fa-44fb-ae36-f334c9bd9d68" (UID: "714c50bd-a6fa-44fb-ae36-f334c9bd9d68"). InnerVolumeSpecName "kube-api-access-sfzxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.334246 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-inventory" (OuterVolumeSpecName: "inventory") pod "714c50bd-a6fa-44fb-ae36-f334c9bd9d68" (UID: "714c50bd-a6fa-44fb-ae36-f334c9bd9d68"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.334884 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "714c50bd-a6fa-44fb-ae36-f334c9bd9d68" (UID: "714c50bd-a6fa-44fb-ae36-f334c9bd9d68"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.406936 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.406974 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfzxw\" (UniqueName: \"kubernetes.io/projected/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-kube-api-access-sfzxw\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.406984 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/714c50bd-a6fa-44fb-ae36-f334c9bd9d68-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.864396 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" event={"ID":"714c50bd-a6fa-44fb-ae36-f334c9bd9d68","Type":"ContainerDied","Data":"f6c189c77300241a55c00e07dd58b94f511c041bd39b40da6ea64bccaa5cb8b8"} Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.864442 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6c189c77300241a55c00e07dd58b94f511c041bd39b40da6ea64bccaa5cb8b8" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.864449 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-pmt2r" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.926530 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg"] Oct 04 04:58:45 crc kubenswrapper[4992]: E1004 04:58:45.926923 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="714c50bd-a6fa-44fb-ae36-f334c9bd9d68" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.926940 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="714c50bd-a6fa-44fb-ae36-f334c9bd9d68" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.927113 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="714c50bd-a6fa-44fb-ae36-f334c9bd9d68" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.927768 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.930142 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.930260 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.930168 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.931397 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 04:58:45 crc kubenswrapper[4992]: I1004 04:58:45.951972 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg"] Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.119646 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.120622 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjbhl\" (UniqueName: \"kubernetes.io/projected/645bddf3-cf63-40fc-b847-40c720339f80-kube-api-access-zjbhl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.120692 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.120714 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.222574 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjbhl\" (UniqueName: \"kubernetes.io/projected/645bddf3-cf63-40fc-b847-40c720339f80-kube-api-access-zjbhl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.222640 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.222670 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.222798 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.228707 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.236888 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.238914 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.240213 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjbhl\" (UniqueName: \"kubernetes.io/projected/645bddf3-cf63-40fc-b847-40c720339f80-kube-api-access-zjbhl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.266441 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.789303 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg"] Oct 04 04:58:46 crc kubenswrapper[4992]: I1004 04:58:46.875656 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" event={"ID":"645bddf3-cf63-40fc-b847-40c720339f80","Type":"ContainerStarted","Data":"d4fcb97618b4d76497308702c262412aa8c65af6cd14da4c5996939f76d4600b"} Oct 04 04:58:47 crc kubenswrapper[4992]: I1004 04:58:47.885863 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" event={"ID":"645bddf3-cf63-40fc-b847-40c720339f80","Type":"ContainerStarted","Data":"7e8b779610b8e657d8bb62966cb2a28b722db1f452102931a1e648949b26da8a"} Oct 04 04:58:47 crc kubenswrapper[4992]: I1004 04:58:47.904721 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" podStartSLOduration=2.4571233599999998 podStartE2EDuration="2.904702987s" podCreationTimestamp="2025-10-04 04:58:45 +0000 UTC" firstStartedPulling="2025-10-04 04:58:46.793931691 +0000 UTC m=+1560.641607149" lastFinishedPulling="2025-10-04 04:58:47.241511308 +0000 UTC m=+1561.089186776" observedRunningTime="2025-10-04 04:58:47.901256885 +0000 UTC m=+1561.748932373" watchObservedRunningTime="2025-10-04 04:58:47.904702987 +0000 UTC m=+1561.752378455" Oct 04 04:59:09 crc kubenswrapper[4992]: I1004 04:59:09.637311 4992 scope.go:117] "RemoveContainer" containerID="b331148fadb0665defb0faacc69aa53d565b349c47baf03f4e851a44fe50069d" Oct 04 04:59:09 crc kubenswrapper[4992]: I1004 04:59:09.666932 4992 scope.go:117] "RemoveContainer" containerID="d13a59a1df114a106065257bdc90f172626d7d877c5190bfb587393a78039c2f" Oct 04 04:59:09 crc kubenswrapper[4992]: I1004 04:59:09.731488 4992 scope.go:117] "RemoveContainer" containerID="0eb64e942f7db33e5f76215eb6d4c90ca70cfc664afcfd27a8ac2410f58e2fed" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.007529 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6h9q5"] Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.012006 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.023148 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6h9q5"] Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.048321 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-utilities\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.048439 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-catalog-content\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.048492 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbz8p\" (UniqueName: \"kubernetes.io/projected/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-kube-api-access-nbz8p\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.150167 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-utilities\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.150226 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-catalog-content\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.150252 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbz8p\" (UniqueName: \"kubernetes.io/projected/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-kube-api-access-nbz8p\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.150644 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-utilities\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.150773 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-catalog-content\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.173138 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbz8p\" (UniqueName: \"kubernetes.io/projected/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-kube-api-access-nbz8p\") pod \"redhat-marketplace-6h9q5\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.338585 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:19 crc kubenswrapper[4992]: I1004 04:59:19.816746 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6h9q5"] Oct 04 04:59:19 crc kubenswrapper[4992]: W1004 04:59:19.819193 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa840562_2ee0_4529_8a9c_79b4e1f4e03c.slice/crio-3860d811f844a7bf96c4052b3061dfee83a4fa032a8585e5ffbd2d394ccbaf35 WatchSource:0}: Error finding container 3860d811f844a7bf96c4052b3061dfee83a4fa032a8585e5ffbd2d394ccbaf35: Status 404 returned error can't find the container with id 3860d811f844a7bf96c4052b3061dfee83a4fa032a8585e5ffbd2d394ccbaf35 Oct 04 04:59:20 crc kubenswrapper[4992]: I1004 04:59:20.169779 4992 generic.go:334] "Generic (PLEG): container finished" podID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerID="3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692" exitCode=0 Oct 04 04:59:20 crc kubenswrapper[4992]: I1004 04:59:20.169848 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6h9q5" event={"ID":"fa840562-2ee0-4529-8a9c-79b4e1f4e03c","Type":"ContainerDied","Data":"3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692"} Oct 04 04:59:20 crc kubenswrapper[4992]: I1004 04:59:20.170064 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6h9q5" event={"ID":"fa840562-2ee0-4529-8a9c-79b4e1f4e03c","Type":"ContainerStarted","Data":"3860d811f844a7bf96c4052b3061dfee83a4fa032a8585e5ffbd2d394ccbaf35"} Oct 04 04:59:22 crc kubenswrapper[4992]: I1004 04:59:22.190617 4992 generic.go:334] "Generic (PLEG): container finished" podID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerID="9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a" exitCode=0 Oct 04 04:59:22 crc kubenswrapper[4992]: I1004 04:59:22.190668 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6h9q5" event={"ID":"fa840562-2ee0-4529-8a9c-79b4e1f4e03c","Type":"ContainerDied","Data":"9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a"} Oct 04 04:59:23 crc kubenswrapper[4992]: I1004 04:59:23.204590 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6h9q5" event={"ID":"fa840562-2ee0-4529-8a9c-79b4e1f4e03c","Type":"ContainerStarted","Data":"da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938"} Oct 04 04:59:23 crc kubenswrapper[4992]: I1004 04:59:23.223758 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6h9q5" podStartSLOduration=2.7603083059999998 podStartE2EDuration="5.223738226s" podCreationTimestamp="2025-10-04 04:59:18 +0000 UTC" firstStartedPulling="2025-10-04 04:59:20.17182688 +0000 UTC m=+1594.019502378" lastFinishedPulling="2025-10-04 04:59:22.63525682 +0000 UTC m=+1596.482932298" observedRunningTime="2025-10-04 04:59:23.220674354 +0000 UTC m=+1597.068349842" watchObservedRunningTime="2025-10-04 04:59:23.223738226 +0000 UTC m=+1597.071413694" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.478274 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-htn47"] Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.482284 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.490763 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-htn47"] Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.610375 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-catalog-content\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.610423 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj9st\" (UniqueName: \"kubernetes.io/projected/f4d904e4-251f-494e-8016-17b1ecdd53ec-kube-api-access-pj9st\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.611069 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-utilities\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.713275 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-catalog-content\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.713353 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj9st\" (UniqueName: \"kubernetes.io/projected/f4d904e4-251f-494e-8016-17b1ecdd53ec-kube-api-access-pj9st\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.713432 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-utilities\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.713845 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-catalog-content\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.714024 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-utilities\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.736204 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj9st\" (UniqueName: \"kubernetes.io/projected/f4d904e4-251f-494e-8016-17b1ecdd53ec-kube-api-access-pj9st\") pod \"certified-operators-htn47\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:26 crc kubenswrapper[4992]: I1004 04:59:26.813864 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:27 crc kubenswrapper[4992]: I1004 04:59:27.314674 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-htn47"] Oct 04 04:59:28 crc kubenswrapper[4992]: I1004 04:59:28.260448 4992 generic.go:334] "Generic (PLEG): container finished" podID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerID="74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8" exitCode=0 Oct 04 04:59:28 crc kubenswrapper[4992]: I1004 04:59:28.260518 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htn47" event={"ID":"f4d904e4-251f-494e-8016-17b1ecdd53ec","Type":"ContainerDied","Data":"74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8"} Oct 04 04:59:28 crc kubenswrapper[4992]: I1004 04:59:28.260750 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htn47" event={"ID":"f4d904e4-251f-494e-8016-17b1ecdd53ec","Type":"ContainerStarted","Data":"ffabb32a22f22ab9d939ab64991a52c448bb4fd629a5a3c0e3f425f7875bc040"} Oct 04 04:59:29 crc kubenswrapper[4992]: I1004 04:59:29.274784 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htn47" event={"ID":"f4d904e4-251f-494e-8016-17b1ecdd53ec","Type":"ContainerStarted","Data":"7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771"} Oct 04 04:59:29 crc kubenswrapper[4992]: I1004 04:59:29.339312 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:29 crc kubenswrapper[4992]: I1004 04:59:29.339729 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:29 crc kubenswrapper[4992]: I1004 04:59:29.390744 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:30 crc kubenswrapper[4992]: I1004 04:59:30.286315 4992 generic.go:334] "Generic (PLEG): container finished" podID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerID="7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771" exitCode=0 Oct 04 04:59:30 crc kubenswrapper[4992]: I1004 04:59:30.286441 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htn47" event={"ID":"f4d904e4-251f-494e-8016-17b1ecdd53ec","Type":"ContainerDied","Data":"7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771"} Oct 04 04:59:30 crc kubenswrapper[4992]: I1004 04:59:30.334957 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:31 crc kubenswrapper[4992]: I1004 04:59:31.659567 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6h9q5"] Oct 04 04:59:32 crc kubenswrapper[4992]: I1004 04:59:32.311102 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htn47" event={"ID":"f4d904e4-251f-494e-8016-17b1ecdd53ec","Type":"ContainerStarted","Data":"e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa"} Oct 04 04:59:32 crc kubenswrapper[4992]: I1004 04:59:32.337396 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-htn47" podStartSLOduration=3.420655787 podStartE2EDuration="6.337375334s" podCreationTimestamp="2025-10-04 04:59:26 +0000 UTC" firstStartedPulling="2025-10-04 04:59:28.26232988 +0000 UTC m=+1602.110005348" lastFinishedPulling="2025-10-04 04:59:31.179049427 +0000 UTC m=+1605.026724895" observedRunningTime="2025-10-04 04:59:32.333932183 +0000 UTC m=+1606.181607661" watchObservedRunningTime="2025-10-04 04:59:32.337375334 +0000 UTC m=+1606.185050802" Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.324710 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6h9q5" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerName="registry-server" containerID="cri-o://da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938" gracePeriod=2 Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.786149 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.855015 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-catalog-content\") pod \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.855297 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-utilities\") pod \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.855383 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbz8p\" (UniqueName: \"kubernetes.io/projected/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-kube-api-access-nbz8p\") pod \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\" (UID: \"fa840562-2ee0-4529-8a9c-79b4e1f4e03c\") " Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.856480 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-utilities" (OuterVolumeSpecName: "utilities") pod "fa840562-2ee0-4529-8a9c-79b4e1f4e03c" (UID: "fa840562-2ee0-4529-8a9c-79b4e1f4e03c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.864521 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-kube-api-access-nbz8p" (OuterVolumeSpecName: "kube-api-access-nbz8p") pod "fa840562-2ee0-4529-8a9c-79b4e1f4e03c" (UID: "fa840562-2ee0-4529-8a9c-79b4e1f4e03c"). InnerVolumeSpecName "kube-api-access-nbz8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.875208 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa840562-2ee0-4529-8a9c-79b4e1f4e03c" (UID: "fa840562-2ee0-4529-8a9c-79b4e1f4e03c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.957455 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.957489 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:33 crc kubenswrapper[4992]: I1004 04:59:33.957500 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbz8p\" (UniqueName: \"kubernetes.io/projected/fa840562-2ee0-4529-8a9c-79b4e1f4e03c-kube-api-access-nbz8p\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.335706 4992 generic.go:334] "Generic (PLEG): container finished" podID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerID="da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938" exitCode=0 Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.335752 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6h9q5" event={"ID":"fa840562-2ee0-4529-8a9c-79b4e1f4e03c","Type":"ContainerDied","Data":"da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938"} Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.335754 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6h9q5" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.335794 4992 scope.go:117] "RemoveContainer" containerID="da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.335779 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6h9q5" event={"ID":"fa840562-2ee0-4529-8a9c-79b4e1f4e03c","Type":"ContainerDied","Data":"3860d811f844a7bf96c4052b3061dfee83a4fa032a8585e5ffbd2d394ccbaf35"} Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.358910 4992 scope.go:117] "RemoveContainer" containerID="9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.370804 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6h9q5"] Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.381050 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6h9q5"] Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.395103 4992 scope.go:117] "RemoveContainer" containerID="3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.428837 4992 scope.go:117] "RemoveContainer" containerID="da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938" Oct 04 04:59:34 crc kubenswrapper[4992]: E1004 04:59:34.429322 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938\": container with ID starting with da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938 not found: ID does not exist" containerID="da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.429381 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938"} err="failed to get container status \"da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938\": rpc error: code = NotFound desc = could not find container \"da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938\": container with ID starting with da81a7f442058571a513fd41c4a0a1be9b6aa3b5bfa80f99caaf1c748a370938 not found: ID does not exist" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.429416 4992 scope.go:117] "RemoveContainer" containerID="9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a" Oct 04 04:59:34 crc kubenswrapper[4992]: E1004 04:59:34.429878 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a\": container with ID starting with 9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a not found: ID does not exist" containerID="9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.429920 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a"} err="failed to get container status \"9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a\": rpc error: code = NotFound desc = could not find container \"9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a\": container with ID starting with 9e7cd46c98550d03bbe7a927e4330653b401d4491cb6662bda8b53e178a5341a not found: ID does not exist" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.429945 4992 scope.go:117] "RemoveContainer" containerID="3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692" Oct 04 04:59:34 crc kubenswrapper[4992]: E1004 04:59:34.430232 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692\": container with ID starting with 3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692 not found: ID does not exist" containerID="3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692" Oct 04 04:59:34 crc kubenswrapper[4992]: I1004 04:59:34.430258 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692"} err="failed to get container status \"3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692\": rpc error: code = NotFound desc = could not find container \"3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692\": container with ID starting with 3c7d77da5ecb9bceb3d97989d04231103f53cb4364eb2cd99da783ddfd44a692 not found: ID does not exist" Oct 04 04:59:36 crc kubenswrapper[4992]: I1004 04:59:36.329633 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" path="/var/lib/kubelet/pods/fa840562-2ee0-4529-8a9c-79b4e1f4e03c/volumes" Oct 04 04:59:36 crc kubenswrapper[4992]: I1004 04:59:36.814331 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:36 crc kubenswrapper[4992]: I1004 04:59:36.814474 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:36 crc kubenswrapper[4992]: I1004 04:59:36.869808 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:37 crc kubenswrapper[4992]: I1004 04:59:37.415953 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:38 crc kubenswrapper[4992]: I1004 04:59:38.256773 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-htn47"] Oct 04 04:59:39 crc kubenswrapper[4992]: I1004 04:59:39.379265 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-htn47" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerName="registry-server" containerID="cri-o://e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa" gracePeriod=2 Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.000248 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.165766 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj9st\" (UniqueName: \"kubernetes.io/projected/f4d904e4-251f-494e-8016-17b1ecdd53ec-kube-api-access-pj9st\") pod \"f4d904e4-251f-494e-8016-17b1ecdd53ec\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.165832 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-catalog-content\") pod \"f4d904e4-251f-494e-8016-17b1ecdd53ec\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.165928 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-utilities\") pod \"f4d904e4-251f-494e-8016-17b1ecdd53ec\" (UID: \"f4d904e4-251f-494e-8016-17b1ecdd53ec\") " Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.167706 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-utilities" (OuterVolumeSpecName: "utilities") pod "f4d904e4-251f-494e-8016-17b1ecdd53ec" (UID: "f4d904e4-251f-494e-8016-17b1ecdd53ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.172447 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4d904e4-251f-494e-8016-17b1ecdd53ec-kube-api-access-pj9st" (OuterVolumeSpecName: "kube-api-access-pj9st") pod "f4d904e4-251f-494e-8016-17b1ecdd53ec" (UID: "f4d904e4-251f-494e-8016-17b1ecdd53ec"). InnerVolumeSpecName "kube-api-access-pj9st". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.268264 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj9st\" (UniqueName: \"kubernetes.io/projected/f4d904e4-251f-494e-8016-17b1ecdd53ec-kube-api-access-pj9st\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.268505 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.393129 4992 generic.go:334] "Generic (PLEG): container finished" podID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerID="e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa" exitCode=0 Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.393145 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4d904e4-251f-494e-8016-17b1ecdd53ec" (UID: "f4d904e4-251f-494e-8016-17b1ecdd53ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.393208 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-htn47" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.393207 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htn47" event={"ID":"f4d904e4-251f-494e-8016-17b1ecdd53ec","Type":"ContainerDied","Data":"e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa"} Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.393331 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-htn47" event={"ID":"f4d904e4-251f-494e-8016-17b1ecdd53ec","Type":"ContainerDied","Data":"ffabb32a22f22ab9d939ab64991a52c448bb4fd629a5a3c0e3f425f7875bc040"} Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.393409 4992 scope.go:117] "RemoveContainer" containerID="e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.424560 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-htn47"] Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.425272 4992 scope.go:117] "RemoveContainer" containerID="7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.432631 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-htn47"] Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.473073 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4d904e4-251f-494e-8016-17b1ecdd53ec-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.480214 4992 scope.go:117] "RemoveContainer" containerID="74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.497759 4992 scope.go:117] "RemoveContainer" containerID="e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa" Oct 04 04:59:40 crc kubenswrapper[4992]: E1004 04:59:40.498232 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa\": container with ID starting with e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa not found: ID does not exist" containerID="e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.498288 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa"} err="failed to get container status \"e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa\": rpc error: code = NotFound desc = could not find container \"e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa\": container with ID starting with e7bcad1a29b7a06ef25c6f394f4190a98cb90bea1463028cdcc09b84b57815aa not found: ID does not exist" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.498310 4992 scope.go:117] "RemoveContainer" containerID="7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771" Oct 04 04:59:40 crc kubenswrapper[4992]: E1004 04:59:40.498873 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771\": container with ID starting with 7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771 not found: ID does not exist" containerID="7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.498914 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771"} err="failed to get container status \"7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771\": rpc error: code = NotFound desc = could not find container \"7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771\": container with ID starting with 7ebd7232d04b818b8d7685b78c272b0dc1e919d847ca27d1dbd9964a123b4771 not found: ID does not exist" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.498942 4992 scope.go:117] "RemoveContainer" containerID="74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8" Oct 04 04:59:40 crc kubenswrapper[4992]: E1004 04:59:40.499518 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8\": container with ID starting with 74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8 not found: ID does not exist" containerID="74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8" Oct 04 04:59:40 crc kubenswrapper[4992]: I1004 04:59:40.499602 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8"} err="failed to get container status \"74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8\": rpc error: code = NotFound desc = could not find container \"74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8\": container with ID starting with 74c227e6a82f4123ece996a9181f75cd41a23d3ab511aaa6687dbf4e5a277de8 not found: ID does not exist" Oct 04 04:59:42 crc kubenswrapper[4992]: I1004 04:59:42.336511 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" path="/var/lib/kubelet/pods/f4d904e4-251f-494e-8016-17b1ecdd53ec/volumes" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.930354 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xnkgb"] Oct 04 04:59:51 crc kubenswrapper[4992]: E1004 04:59:51.931449 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerName="extract-utilities" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.931468 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerName="extract-utilities" Oct 04 04:59:51 crc kubenswrapper[4992]: E1004 04:59:51.931479 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerName="registry-server" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.931486 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerName="registry-server" Oct 04 04:59:51 crc kubenswrapper[4992]: E1004 04:59:51.931502 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerName="extract-content" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.931510 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerName="extract-content" Oct 04 04:59:51 crc kubenswrapper[4992]: E1004 04:59:51.931525 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerName="extract-content" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.931532 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerName="extract-content" Oct 04 04:59:51 crc kubenswrapper[4992]: E1004 04:59:51.931555 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerName="registry-server" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.931564 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerName="registry-server" Oct 04 04:59:51 crc kubenswrapper[4992]: E1004 04:59:51.931582 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerName="extract-utilities" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.931590 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerName="extract-utilities" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.931819 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4d904e4-251f-494e-8016-17b1ecdd53ec" containerName="registry-server" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.931837 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa840562-2ee0-4529-8a9c-79b4e1f4e03c" containerName="registry-server" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.933466 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:51 crc kubenswrapper[4992]: I1004 04:59:51.947022 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xnkgb"] Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.101510 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-utilities\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.101666 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td4bn\" (UniqueName: \"kubernetes.io/projected/6a226c16-017e-48bf-9808-e7b6dc34a826-kube-api-access-td4bn\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.101727 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-catalog-content\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.204425 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-utilities\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.203861 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-utilities\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.204926 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td4bn\" (UniqueName: \"kubernetes.io/projected/6a226c16-017e-48bf-9808-e7b6dc34a826-kube-api-access-td4bn\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.205014 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-catalog-content\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.205384 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-catalog-content\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.228461 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td4bn\" (UniqueName: \"kubernetes.io/projected/6a226c16-017e-48bf-9808-e7b6dc34a826-kube-api-access-td4bn\") pod \"community-operators-xnkgb\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.261717 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnkgb" Oct 04 04:59:52 crc kubenswrapper[4992]: I1004 04:59:52.798972 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xnkgb"] Oct 04 04:59:53 crc kubenswrapper[4992]: I1004 04:59:53.511064 4992 generic.go:334] "Generic (PLEG): container finished" podID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerID="4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8" exitCode=0 Oct 04 04:59:53 crc kubenswrapper[4992]: I1004 04:59:53.511159 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnkgb" event={"ID":"6a226c16-017e-48bf-9808-e7b6dc34a826","Type":"ContainerDied","Data":"4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8"} Oct 04 04:59:53 crc kubenswrapper[4992]: I1004 04:59:53.511348 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnkgb" event={"ID":"6a226c16-017e-48bf-9808-e7b6dc34a826","Type":"ContainerStarted","Data":"28b345a8b1845df9432dc36f221046404c1f0d2508db0010326b7104462d018c"} Oct 04 04:59:55 crc kubenswrapper[4992]: I1004 04:59:55.532668 4992 generic.go:334] "Generic (PLEG): container finished" podID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerID="a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac" exitCode=0 Oct 04 04:59:55 crc kubenswrapper[4992]: I1004 04:59:55.532782 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnkgb" event={"ID":"6a226c16-017e-48bf-9808-e7b6dc34a826","Type":"ContainerDied","Data":"a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac"} Oct 04 04:59:56 crc kubenswrapper[4992]: I1004 04:59:56.542457 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnkgb" event={"ID":"6a226c16-017e-48bf-9808-e7b6dc34a826","Type":"ContainerStarted","Data":"d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536"} Oct 04 04:59:56 crc kubenswrapper[4992]: I1004 04:59:56.562113 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xnkgb" podStartSLOduration=3.085173337 podStartE2EDuration="5.562094104s" podCreationTimestamp="2025-10-04 04:59:51 +0000 UTC" firstStartedPulling="2025-10-04 04:59:53.515045038 +0000 UTC m=+1627.362720506" lastFinishedPulling="2025-10-04 04:59:55.991965805 +0000 UTC m=+1629.839641273" observedRunningTime="2025-10-04 04:59:56.558647682 +0000 UTC m=+1630.406323170" watchObservedRunningTime="2025-10-04 04:59:56.562094104 +0000 UTC m=+1630.409769572" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.161845 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7"] Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.163997 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.166669 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.167661 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.173303 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7"] Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.259614 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-config-volume\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.259790 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8cm5\" (UniqueName: \"kubernetes.io/projected/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-kube-api-access-b8cm5\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.260176 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-secret-volume\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.361353 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-secret-volume\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.361448 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-config-volume\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.361500 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8cm5\" (UniqueName: \"kubernetes.io/projected/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-kube-api-access-b8cm5\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.362788 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-config-volume\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.369830 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-secret-volume\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.389769 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8cm5\" (UniqueName: \"kubernetes.io/projected/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-kube-api-access-b8cm5\") pod \"collect-profiles-29325900-zf8n7\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.486925 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:00 crc kubenswrapper[4992]: I1004 05:00:00.973810 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7"] Oct 04 05:00:01 crc kubenswrapper[4992]: I1004 05:00:01.601736 4992 generic.go:334] "Generic (PLEG): container finished" podID="f9d992b6-7a89-4ebe-b5e4-b845019b26f4" containerID="a057a37fcbe912bed8e156b72be425604726ffaa272d04c3ab309fb33f57c3c1" exitCode=0 Oct 04 05:00:01 crc kubenswrapper[4992]: I1004 05:00:01.601836 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" event={"ID":"f9d992b6-7a89-4ebe-b5e4-b845019b26f4","Type":"ContainerDied","Data":"a057a37fcbe912bed8e156b72be425604726ffaa272d04c3ab309fb33f57c3c1"} Oct 04 05:00:01 crc kubenswrapper[4992]: I1004 05:00:01.602084 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" event={"ID":"f9d992b6-7a89-4ebe-b5e4-b845019b26f4","Type":"ContainerStarted","Data":"3dbf9c8843313a020447620746c5d95ab1474e33c0505d745f6103d60ae0b7a3"} Oct 04 05:00:02 crc kubenswrapper[4992]: I1004 05:00:02.262786 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xnkgb" Oct 04 05:00:02 crc kubenswrapper[4992]: I1004 05:00:02.263146 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xnkgb" Oct 04 05:00:02 crc kubenswrapper[4992]: I1004 05:00:02.317283 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xnkgb" Oct 04 05:00:02 crc kubenswrapper[4992]: I1004 05:00:02.667148 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xnkgb" Oct 04 05:00:02 crc kubenswrapper[4992]: I1004 05:00:02.729653 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xnkgb"] Oct 04 05:00:02 crc kubenswrapper[4992]: I1004 05:00:02.977167 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.115672 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8cm5\" (UniqueName: \"kubernetes.io/projected/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-kube-api-access-b8cm5\") pod \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.115803 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-config-volume\") pod \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.116081 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-secret-volume\") pod \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\" (UID: \"f9d992b6-7a89-4ebe-b5e4-b845019b26f4\") " Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.117415 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-config-volume" (OuterVolumeSpecName: "config-volume") pod "f9d992b6-7a89-4ebe-b5e4-b845019b26f4" (UID: "f9d992b6-7a89-4ebe-b5e4-b845019b26f4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.122486 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f9d992b6-7a89-4ebe-b5e4-b845019b26f4" (UID: "f9d992b6-7a89-4ebe-b5e4-b845019b26f4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.122700 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-kube-api-access-b8cm5" (OuterVolumeSpecName: "kube-api-access-b8cm5") pod "f9d992b6-7a89-4ebe-b5e4-b845019b26f4" (UID: "f9d992b6-7a89-4ebe-b5e4-b845019b26f4"). InnerVolumeSpecName "kube-api-access-b8cm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.218664 4992 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.218700 4992 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.218710 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8cm5\" (UniqueName: \"kubernetes.io/projected/f9d992b6-7a89-4ebe-b5e4-b845019b26f4-kube-api-access-b8cm5\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.624474 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.624477 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7" event={"ID":"f9d992b6-7a89-4ebe-b5e4-b845019b26f4","Type":"ContainerDied","Data":"3dbf9c8843313a020447620746c5d95ab1474e33c0505d745f6103d60ae0b7a3"} Oct 04 05:00:03 crc kubenswrapper[4992]: I1004 05:00:03.624872 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dbf9c8843313a020447620746c5d95ab1474e33c0505d745f6103d60ae0b7a3" Oct 04 05:00:04 crc kubenswrapper[4992]: I1004 05:00:04.635200 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xnkgb" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerName="registry-server" containerID="cri-o://d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536" gracePeriod=2 Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.119836 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnkgb" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.261053 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-utilities\") pod \"6a226c16-017e-48bf-9808-e7b6dc34a826\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.261153 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td4bn\" (UniqueName: \"kubernetes.io/projected/6a226c16-017e-48bf-9808-e7b6dc34a826-kube-api-access-td4bn\") pod \"6a226c16-017e-48bf-9808-e7b6dc34a826\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.261296 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-catalog-content\") pod \"6a226c16-017e-48bf-9808-e7b6dc34a826\" (UID: \"6a226c16-017e-48bf-9808-e7b6dc34a826\") " Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.262139 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-utilities" (OuterVolumeSpecName: "utilities") pod "6a226c16-017e-48bf-9808-e7b6dc34a826" (UID: "6a226c16-017e-48bf-9808-e7b6dc34a826"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.267508 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a226c16-017e-48bf-9808-e7b6dc34a826-kube-api-access-td4bn" (OuterVolumeSpecName: "kube-api-access-td4bn") pod "6a226c16-017e-48bf-9808-e7b6dc34a826" (UID: "6a226c16-017e-48bf-9808-e7b6dc34a826"). InnerVolumeSpecName "kube-api-access-td4bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.312113 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a226c16-017e-48bf-9808-e7b6dc34a826" (UID: "6a226c16-017e-48bf-9808-e7b6dc34a826"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.364408 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.364444 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a226c16-017e-48bf-9808-e7b6dc34a826-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.364457 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td4bn\" (UniqueName: \"kubernetes.io/projected/6a226c16-017e-48bf-9808-e7b6dc34a826-kube-api-access-td4bn\") on node \"crc\" DevicePath \"\"" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.648140 4992 generic.go:334] "Generic (PLEG): container finished" podID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerID="d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536" exitCode=0 Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.648191 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnkgb" event={"ID":"6a226c16-017e-48bf-9808-e7b6dc34a826","Type":"ContainerDied","Data":"d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536"} Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.648219 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xnkgb" event={"ID":"6a226c16-017e-48bf-9808-e7b6dc34a826","Type":"ContainerDied","Data":"28b345a8b1845df9432dc36f221046404c1f0d2508db0010326b7104462d018c"} Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.648241 4992 scope.go:117] "RemoveContainer" containerID="d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.648492 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xnkgb" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.683538 4992 scope.go:117] "RemoveContainer" containerID="a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.687827 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xnkgb"] Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.698755 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xnkgb"] Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.716922 4992 scope.go:117] "RemoveContainer" containerID="4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.756542 4992 scope.go:117] "RemoveContainer" containerID="d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536" Oct 04 05:00:05 crc kubenswrapper[4992]: E1004 05:00:05.757022 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536\": container with ID starting with d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536 not found: ID does not exist" containerID="d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.757062 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536"} err="failed to get container status \"d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536\": rpc error: code = NotFound desc = could not find container \"d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536\": container with ID starting with d7725ef4a7d8aa2b399135890c6b9d1cea9b1775f046e926178285c3e2a5e536 not found: ID does not exist" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.757089 4992 scope.go:117] "RemoveContainer" containerID="a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac" Oct 04 05:00:05 crc kubenswrapper[4992]: E1004 05:00:05.757370 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac\": container with ID starting with a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac not found: ID does not exist" containerID="a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.757400 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac"} err="failed to get container status \"a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac\": rpc error: code = NotFound desc = could not find container \"a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac\": container with ID starting with a9397e23a53f3f44adf987368261ccfc5628e3e0a9623975146d155598d71cac not found: ID does not exist" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.757421 4992 scope.go:117] "RemoveContainer" containerID="4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8" Oct 04 05:00:05 crc kubenswrapper[4992]: E1004 05:00:05.757854 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8\": container with ID starting with 4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8 not found: ID does not exist" containerID="4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8" Oct 04 05:00:05 crc kubenswrapper[4992]: I1004 05:00:05.757876 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8"} err="failed to get container status \"4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8\": rpc error: code = NotFound desc = could not find container \"4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8\": container with ID starting with 4079bfb15499aa2360599a43cd5e5f715ed50fef1d702f36b3f673bbf3a4dbb8 not found: ID does not exist" Oct 04 05:00:06 crc kubenswrapper[4992]: I1004 05:00:06.330681 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" path="/var/lib/kubelet/pods/6a226c16-017e-48bf-9808-e7b6dc34a826/volumes" Oct 04 05:00:20 crc kubenswrapper[4992]: I1004 05:00:20.894858 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:00:20 crc kubenswrapper[4992]: I1004 05:00:20.895613 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:00:50 crc kubenswrapper[4992]: I1004 05:00:50.894458 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:00:50 crc kubenswrapper[4992]: I1004 05:00:50.895022 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.166204 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29325901-dvngf"] Oct 04 05:01:00 crc kubenswrapper[4992]: E1004 05:01:00.167459 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerName="extract-utilities" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.167478 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerName="extract-utilities" Oct 04 05:01:00 crc kubenswrapper[4992]: E1004 05:01:00.167517 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.167524 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4992]: E1004 05:01:00.167534 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9d992b6-7a89-4ebe-b5e4-b845019b26f4" containerName="collect-profiles" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.167540 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9d992b6-7a89-4ebe-b5e4-b845019b26f4" containerName="collect-profiles" Oct 04 05:01:00 crc kubenswrapper[4992]: E1004 05:01:00.167563 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerName="extract-content" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.167569 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerName="extract-content" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.167759 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a226c16-017e-48bf-9808-e7b6dc34a826" containerName="registry-server" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.167777 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9d992b6-7a89-4ebe-b5e4-b845019b26f4" containerName="collect-profiles" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.168627 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.176904 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325901-dvngf"] Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.190790 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-fernet-keys\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.190874 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-combined-ca-bundle\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.191399 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb7lm\" (UniqueName: \"kubernetes.io/projected/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-kube-api-access-rb7lm\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.191908 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-config-data\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.294583 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb7lm\" (UniqueName: \"kubernetes.io/projected/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-kube-api-access-rb7lm\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.294715 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-config-data\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.294751 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-fernet-keys\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.294777 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-combined-ca-bundle\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.302667 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-fernet-keys\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.302997 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-config-data\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.306645 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-combined-ca-bundle\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.312186 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb7lm\" (UniqueName: \"kubernetes.io/projected/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-kube-api-access-rb7lm\") pod \"keystone-cron-29325901-dvngf\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.498937 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:00 crc kubenswrapper[4992]: I1004 05:01:00.954686 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29325901-dvngf"] Oct 04 05:01:01 crc kubenswrapper[4992]: I1004 05:01:01.159639 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-dvngf" event={"ID":"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f","Type":"ContainerStarted","Data":"3d8d0d603d3e848e0fbc01d0155cf119cc75e42c7aa00bbe371752c024403d46"} Oct 04 05:01:02 crc kubenswrapper[4992]: I1004 05:01:02.170947 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-dvngf" event={"ID":"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f","Type":"ContainerStarted","Data":"9dbc502a333dd558fcd86a916b8b55a015eda50813f00800d66896fb4cb86e49"} Oct 04 05:01:02 crc kubenswrapper[4992]: I1004 05:01:02.209995 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29325901-dvngf" podStartSLOduration=2.209975784 podStartE2EDuration="2.209975784s" podCreationTimestamp="2025-10-04 05:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:01:02.204894299 +0000 UTC m=+1696.052569777" watchObservedRunningTime="2025-10-04 05:01:02.209975784 +0000 UTC m=+1696.057651252" Oct 04 05:01:04 crc kubenswrapper[4992]: I1004 05:01:04.190731 4992 generic.go:334] "Generic (PLEG): container finished" podID="2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f" containerID="9dbc502a333dd558fcd86a916b8b55a015eda50813f00800d66896fb4cb86e49" exitCode=0 Oct 04 05:01:04 crc kubenswrapper[4992]: I1004 05:01:04.190822 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-dvngf" event={"ID":"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f","Type":"ContainerDied","Data":"9dbc502a333dd558fcd86a916b8b55a015eda50813f00800d66896fb4cb86e49"} Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.552183 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.691623 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb7lm\" (UniqueName: \"kubernetes.io/projected/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-kube-api-access-rb7lm\") pod \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.691724 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-combined-ca-bundle\") pod \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.691858 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-config-data\") pod \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.691983 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-fernet-keys\") pod \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\" (UID: \"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f\") " Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.709463 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f" (UID: "2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.709608 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-kube-api-access-rb7lm" (OuterVolumeSpecName: "kube-api-access-rb7lm") pod "2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f" (UID: "2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f"). InnerVolumeSpecName "kube-api-access-rb7lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.729482 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f" (UID: "2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.761099 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-config-data" (OuterVolumeSpecName: "config-data") pod "2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f" (UID: "2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.794252 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb7lm\" (UniqueName: \"kubernetes.io/projected/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-kube-api-access-rb7lm\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.794288 4992 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.794298 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:05 crc kubenswrapper[4992]: I1004 05:01:05.794307 4992 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 05:01:06 crc kubenswrapper[4992]: I1004 05:01:06.215675 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29325901-dvngf" event={"ID":"2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f","Type":"ContainerDied","Data":"3d8d0d603d3e848e0fbc01d0155cf119cc75e42c7aa00bbe371752c024403d46"} Oct 04 05:01:06 crc kubenswrapper[4992]: I1004 05:01:06.215717 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d8d0d603d3e848e0fbc01d0155cf119cc75e42c7aa00bbe371752c024403d46" Oct 04 05:01:06 crc kubenswrapper[4992]: I1004 05:01:06.215770 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29325901-dvngf" Oct 04 05:01:20 crc kubenswrapper[4992]: I1004 05:01:20.894042 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:01:20 crc kubenswrapper[4992]: I1004 05:01:20.894783 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:01:20 crc kubenswrapper[4992]: I1004 05:01:20.894838 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 05:01:20 crc kubenswrapper[4992]: I1004 05:01:20.895584 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:01:20 crc kubenswrapper[4992]: I1004 05:01:20.895645 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" gracePeriod=600 Oct 04 05:01:21 crc kubenswrapper[4992]: E1004 05:01:21.044662 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:01:21 crc kubenswrapper[4992]: I1004 05:01:21.352025 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" exitCode=0 Oct 04 05:01:21 crc kubenswrapper[4992]: I1004 05:01:21.352093 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817"} Oct 04 05:01:21 crc kubenswrapper[4992]: I1004 05:01:21.352446 4992 scope.go:117] "RemoveContainer" containerID="b7858b04a08dbf3c2f36756b5a0b3054268a106fb6be0797a45494ff8fb3a8ad" Oct 04 05:01:21 crc kubenswrapper[4992]: I1004 05:01:21.353050 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:01:21 crc kubenswrapper[4992]: E1004 05:01:21.353305 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:01:37 crc kubenswrapper[4992]: I1004 05:01:37.320056 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:01:37 crc kubenswrapper[4992]: E1004 05:01:37.321170 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:01:48 crc kubenswrapper[4992]: I1004 05:01:48.320177 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:01:48 crc kubenswrapper[4992]: E1004 05:01:48.321222 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:02:00 crc kubenswrapper[4992]: I1004 05:02:00.319185 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:02:00 crc kubenswrapper[4992]: E1004 05:02:00.319942 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:02:09 crc kubenswrapper[4992]: I1004 05:02:09.997778 4992 scope.go:117] "RemoveContainer" containerID="472db47216833d39a4fb8cca0b5a138ac1238adcf1e32c55b306b33fb9eaf815" Oct 04 05:02:10 crc kubenswrapper[4992]: I1004 05:02:10.028301 4992 scope.go:117] "RemoveContainer" containerID="c077359219ffa989eb527dd2ee5349d788e958b6de39cc45c31f6caa98e8963e" Oct 04 05:02:10 crc kubenswrapper[4992]: I1004 05:02:10.051960 4992 scope.go:117] "RemoveContainer" containerID="72584ab7b963251881167fb87357ffd30791de1deb7ec50da7e5ea3e1cfaa650" Oct 04 05:02:10 crc kubenswrapper[4992]: I1004 05:02:10.075826 4992 scope.go:117] "RemoveContainer" containerID="2a2358f1c9f9d51ebe58d267edcafd4b0425263be03c3045909384bceca080a5" Oct 04 05:02:10 crc kubenswrapper[4992]: I1004 05:02:10.100819 4992 scope.go:117] "RemoveContainer" containerID="fbc115b24de9a1455b159445670a382c881f3da18b8179226a8111efe1ea33c7" Oct 04 05:02:10 crc kubenswrapper[4992]: I1004 05:02:10.127024 4992 scope.go:117] "RemoveContainer" containerID="21bfee8250581a79c116d16342f7b1fccf51ec429e7927d8deea72270c7f5193" Oct 04 05:02:10 crc kubenswrapper[4992]: I1004 05:02:10.187242 4992 scope.go:117] "RemoveContainer" containerID="a39692110716d57b5deebbbe8249739a6c1141c43fdd884a7dba8e7a956b4114" Oct 04 05:02:10 crc kubenswrapper[4992]: I1004 05:02:10.205981 4992 scope.go:117] "RemoveContainer" containerID="77515990fc208dfe3a4ac0806b9984f73fe631843eec21b537591e8db1dffa51" Oct 04 05:02:14 crc kubenswrapper[4992]: I1004 05:02:14.327291 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:02:14 crc kubenswrapper[4992]: E1004 05:02:14.327879 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:02:20 crc kubenswrapper[4992]: I1004 05:02:20.928120 4992 generic.go:334] "Generic (PLEG): container finished" podID="645bddf3-cf63-40fc-b847-40c720339f80" containerID="7e8b779610b8e657d8bb62966cb2a28b722db1f452102931a1e648949b26da8a" exitCode=0 Oct 04 05:02:20 crc kubenswrapper[4992]: I1004 05:02:20.928199 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" event={"ID":"645bddf3-cf63-40fc-b847-40c720339f80","Type":"ContainerDied","Data":"7e8b779610b8e657d8bb62966cb2a28b722db1f452102931a1e648949b26da8a"} Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.317616 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.454964 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-inventory\") pod \"645bddf3-cf63-40fc-b847-40c720339f80\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.455528 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-ssh-key\") pod \"645bddf3-cf63-40fc-b847-40c720339f80\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.455573 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-bootstrap-combined-ca-bundle\") pod \"645bddf3-cf63-40fc-b847-40c720339f80\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.455626 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjbhl\" (UniqueName: \"kubernetes.io/projected/645bddf3-cf63-40fc-b847-40c720339f80-kube-api-access-zjbhl\") pod \"645bddf3-cf63-40fc-b847-40c720339f80\" (UID: \"645bddf3-cf63-40fc-b847-40c720339f80\") " Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.461567 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/645bddf3-cf63-40fc-b847-40c720339f80-kube-api-access-zjbhl" (OuterVolumeSpecName: "kube-api-access-zjbhl") pod "645bddf3-cf63-40fc-b847-40c720339f80" (UID: "645bddf3-cf63-40fc-b847-40c720339f80"). InnerVolumeSpecName "kube-api-access-zjbhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.462339 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "645bddf3-cf63-40fc-b847-40c720339f80" (UID: "645bddf3-cf63-40fc-b847-40c720339f80"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.483897 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "645bddf3-cf63-40fc-b847-40c720339f80" (UID: "645bddf3-cf63-40fc-b847-40c720339f80"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.485255 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-inventory" (OuterVolumeSpecName: "inventory") pod "645bddf3-cf63-40fc-b847-40c720339f80" (UID: "645bddf3-cf63-40fc-b847-40c720339f80"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.558383 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.558426 4992 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.558441 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjbhl\" (UniqueName: \"kubernetes.io/projected/645bddf3-cf63-40fc-b847-40c720339f80-kube-api-access-zjbhl\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.558456 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/645bddf3-cf63-40fc-b847-40c720339f80-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.947386 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" event={"ID":"645bddf3-cf63-40fc-b847-40c720339f80","Type":"ContainerDied","Data":"d4fcb97618b4d76497308702c262412aa8c65af6cd14da4c5996939f76d4600b"} Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.947424 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4fcb97618b4d76497308702c262412aa8c65af6cd14da4c5996939f76d4600b" Oct 04 05:02:22 crc kubenswrapper[4992]: I1004 05:02:22.947446 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.024700 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c"] Oct 04 05:02:23 crc kubenswrapper[4992]: E1004 05:02:23.025269 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f" containerName="keystone-cron" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.025292 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f" containerName="keystone-cron" Oct 04 05:02:23 crc kubenswrapper[4992]: E1004 05:02:23.025316 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="645bddf3-cf63-40fc-b847-40c720339f80" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.025327 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="645bddf3-cf63-40fc-b847-40c720339f80" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.025583 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="645bddf3-cf63-40fc-b847-40c720339f80" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.025613 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f" containerName="keystone-cron" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.026352 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.028747 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.029016 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.029370 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.030152 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.035334 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c"] Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.168814 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.169031 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctwd8\" (UniqueName: \"kubernetes.io/projected/886d41dd-15e8-48ce-bbe9-f8df87efabba-kube-api-access-ctwd8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.169109 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.270913 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.271079 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctwd8\" (UniqueName: \"kubernetes.io/projected/886d41dd-15e8-48ce-bbe9-f8df87efabba-kube-api-access-ctwd8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.271119 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.274916 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.280304 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.293432 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctwd8\" (UniqueName: \"kubernetes.io/projected/886d41dd-15e8-48ce-bbe9-f8df87efabba-kube-api-access-ctwd8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-cr45c\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.344109 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.840775 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c"] Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.850152 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:02:23 crc kubenswrapper[4992]: I1004 05:02:23.958994 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" event={"ID":"886d41dd-15e8-48ce-bbe9-f8df87efabba","Type":"ContainerStarted","Data":"e4249e9dd8d558a676adeba7e13bea7d435208dd76258e13b31a6de7bd3e2d81"} Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.042009 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-f2flr"] Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.054251 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-4c65b"] Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.063891 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-tqxxk"] Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.071931 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-4c65b"] Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.080329 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-tqxxk"] Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.088736 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-f2flr"] Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.331326 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3365eb93-bd5e-4dee-9386-1382ee1c952e" path="/var/lib/kubelet/pods/3365eb93-bd5e-4dee-9386-1382ee1c952e/volumes" Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.332133 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a21eec04-f852-4e59-ba65-c40a22393706" path="/var/lib/kubelet/pods/a21eec04-f852-4e59-ba65-c40a22393706/volumes" Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.332759 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dedeadd4-7b3f-433c-b310-be8f83cbce0b" path="/var/lib/kubelet/pods/dedeadd4-7b3f-433c-b310-be8f83cbce0b/volumes" Oct 04 05:02:24 crc kubenswrapper[4992]: I1004 05:02:24.969653 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" event={"ID":"886d41dd-15e8-48ce-bbe9-f8df87efabba","Type":"ContainerStarted","Data":"b51edca6b10bdd6085cec62bd22cbea873c2d07971a951ed90eda2c5f2545090"} Oct 04 05:02:25 crc kubenswrapper[4992]: I1004 05:02:25.005622 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" podStartSLOduration=1.301923028 podStartE2EDuration="2.005598238s" podCreationTimestamp="2025-10-04 05:02:23 +0000 UTC" firstStartedPulling="2025-10-04 05:02:23.849258381 +0000 UTC m=+1777.696933849" lastFinishedPulling="2025-10-04 05:02:24.552933591 +0000 UTC m=+1778.400609059" observedRunningTime="2025-10-04 05:02:24.983749477 +0000 UTC m=+1778.831424945" watchObservedRunningTime="2025-10-04 05:02:25.005598238 +0000 UTC m=+1778.853273706" Oct 04 05:02:27 crc kubenswrapper[4992]: I1004 05:02:27.319240 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:02:27 crc kubenswrapper[4992]: E1004 05:02:27.320120 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:02:33 crc kubenswrapper[4992]: I1004 05:02:33.039814 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e810-account-create-bfskd"] Oct 04 05:02:33 crc kubenswrapper[4992]: I1004 05:02:33.053067 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9169-account-create-dpkj4"] Oct 04 05:02:33 crc kubenswrapper[4992]: I1004 05:02:33.062842 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e810-account-create-bfskd"] Oct 04 05:02:33 crc kubenswrapper[4992]: I1004 05:02:33.072913 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9169-account-create-dpkj4"] Oct 04 05:02:34 crc kubenswrapper[4992]: I1004 05:02:34.027233 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-eff6-account-create-sz54v"] Oct 04 05:02:34 crc kubenswrapper[4992]: I1004 05:02:34.036140 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-eff6-account-create-sz54v"] Oct 04 05:02:34 crc kubenswrapper[4992]: I1004 05:02:34.331131 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b8c7673-09e9-4b88-9970-5a3a01eac6a3" path="/var/lib/kubelet/pods/2b8c7673-09e9-4b88-9970-5a3a01eac6a3/volumes" Oct 04 05:02:34 crc kubenswrapper[4992]: I1004 05:02:34.331869 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81a290cb-94bf-41ad-87bf-ccc8a713e03d" path="/var/lib/kubelet/pods/81a290cb-94bf-41ad-87bf-ccc8a713e03d/volumes" Oct 04 05:02:34 crc kubenswrapper[4992]: I1004 05:02:34.332386 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd465ca0-c12b-4f3c-8de2-377ac6d81f89" path="/var/lib/kubelet/pods/dd465ca0-c12b-4f3c-8de2-377ac6d81f89/volumes" Oct 04 05:02:39 crc kubenswrapper[4992]: I1004 05:02:39.319682 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:02:39 crc kubenswrapper[4992]: E1004 05:02:39.320603 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:02:40 crc kubenswrapper[4992]: I1004 05:02:40.029930 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-hrlth"] Oct 04 05:02:40 crc kubenswrapper[4992]: I1004 05:02:40.039164 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-lwcgl"] Oct 04 05:02:40 crc kubenswrapper[4992]: I1004 05:02:40.050064 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-lwcgl"] Oct 04 05:02:40 crc kubenswrapper[4992]: I1004 05:02:40.058609 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-hrlth"] Oct 04 05:02:40 crc kubenswrapper[4992]: I1004 05:02:40.330159 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44" path="/var/lib/kubelet/pods/291bdbc3-5eca-4ab9-8666-3a2b0e5ecd44/volumes" Oct 04 05:02:40 crc kubenswrapper[4992]: I1004 05:02:40.330785 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c3904c3-0dc6-46b0-b152-688ce48cda67" path="/var/lib/kubelet/pods/9c3904c3-0dc6-46b0-b152-688ce48cda67/volumes" Oct 04 05:02:42 crc kubenswrapper[4992]: I1004 05:02:42.023737 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-2z2v2"] Oct 04 05:02:42 crc kubenswrapper[4992]: I1004 05:02:42.032421 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-2z2v2"] Oct 04 05:02:42 crc kubenswrapper[4992]: I1004 05:02:42.330477 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99e52292-79e5-41bd-b360-40002e564a01" path="/var/lib/kubelet/pods/99e52292-79e5-41bd-b360-40002e564a01/volumes" Oct 04 05:02:51 crc kubenswrapper[4992]: I1004 05:02:51.319623 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:02:51 crc kubenswrapper[4992]: E1004 05:02:51.320609 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:02:56 crc kubenswrapper[4992]: I1004 05:02:56.040566 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-204f-account-create-wpsgm"] Oct 04 05:02:56 crc kubenswrapper[4992]: I1004 05:02:56.052527 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-095f-account-create-vdc9q"] Oct 04 05:02:56 crc kubenswrapper[4992]: I1004 05:02:56.060965 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-204f-account-create-wpsgm"] Oct 04 05:02:56 crc kubenswrapper[4992]: I1004 05:02:56.068207 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-095f-account-create-vdc9q"] Oct 04 05:02:56 crc kubenswrapper[4992]: I1004 05:02:56.331736 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="718cab8d-53e1-4d02-940a-4b2882070fa6" path="/var/lib/kubelet/pods/718cab8d-53e1-4d02-940a-4b2882070fa6/volumes" Oct 04 05:02:56 crc kubenswrapper[4992]: I1004 05:02:56.332288 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd" path="/var/lib/kubelet/pods/7b4d3a94-3c31-49a6-9f0b-4b6f7cd227dd/volumes" Oct 04 05:03:00 crc kubenswrapper[4992]: I1004 05:03:00.029556 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-22q9n"] Oct 04 05:03:00 crc kubenswrapper[4992]: I1004 05:03:00.039431 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-22q9n"] Oct 04 05:03:00 crc kubenswrapper[4992]: I1004 05:03:00.049660 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-cc11-account-create-svpcv"] Oct 04 05:03:00 crc kubenswrapper[4992]: I1004 05:03:00.056613 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-cc11-account-create-svpcv"] Oct 04 05:03:00 crc kubenswrapper[4992]: I1004 05:03:00.333678 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b1cd82a-9ceb-4a24-937b-b4917eaa473a" path="/var/lib/kubelet/pods/2b1cd82a-9ceb-4a24-937b-b4917eaa473a/volumes" Oct 04 05:03:00 crc kubenswrapper[4992]: I1004 05:03:00.334235 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5cba52e-0e38-4e46-98a6-4c08da5bd709" path="/var/lib/kubelet/pods/d5cba52e-0e38-4e46-98a6-4c08da5bd709/volumes" Oct 04 05:03:04 crc kubenswrapper[4992]: I1004 05:03:04.327582 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:03:04 crc kubenswrapper[4992]: E1004 05:03:04.328204 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:03:05 crc kubenswrapper[4992]: I1004 05:03:05.029407 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-chkks"] Oct 04 05:03:05 crc kubenswrapper[4992]: I1004 05:03:05.039273 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-chkks"] Oct 04 05:03:06 crc kubenswrapper[4992]: I1004 05:03:06.334389 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea89995-b28e-471b-8d9e-6f97d93fe073" path="/var/lib/kubelet/pods/2ea89995-b28e-471b-8d9e-6f97d93fe073/volumes" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.307614 4992 scope.go:117] "RemoveContainer" containerID="4419ca98afc2b0d3056a2de9cd37f6807c07a3755f066da6d965a09614690f53" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.341676 4992 scope.go:117] "RemoveContainer" containerID="c68804cb71e80e3632d31b028a7a4b55259bcd8520ba5ba9d5efe980650a0d80" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.383767 4992 scope.go:117] "RemoveContainer" containerID="451f9b9f959f53d1ddba9019564ba6b46acf04f31cc977d4ab8d130873fc7056" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.440807 4992 scope.go:117] "RemoveContainer" containerID="c540711e4eb3ba9d17845bc1780e6a24be38964532fb2530c93b2cc32f1cbbbe" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.528688 4992 scope.go:117] "RemoveContainer" containerID="dcd3dbfe58eb58206f09ba1f68da07606ec29892fa372bcf976bda70525664e2" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.550181 4992 scope.go:117] "RemoveContainer" containerID="f5ef9c36afc2be9c54a5499cc20ebbfb57ef5a929e3b1842919345cf22c80c9a" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.594973 4992 scope.go:117] "RemoveContainer" containerID="e067dd9c744fda9f058e838f00a3439d2616d1a94f18522d82fdf28c9eae8e53" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.621043 4992 scope.go:117] "RemoveContainer" containerID="4e88a53ee59cfa6d83eee2666d49888fe8b2437146c38b633dffcaf41a43e83f" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.650588 4992 scope.go:117] "RemoveContainer" containerID="a42fa51e630aa00f894fdc19ce2157d48d9ee5d7f77e4e4c756543c343c66f00" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.682543 4992 scope.go:117] "RemoveContainer" containerID="253c204ba74e7b725c88de437bc908f2df7f5c962d8532e5526e9dbc9693f034" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.708772 4992 scope.go:117] "RemoveContainer" containerID="96c86b7166336c91921d9a11958638b6d887892378a8b31751fb5e321561b55a" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.738596 4992 scope.go:117] "RemoveContainer" containerID="f76dd8f975bddf7ce304c807b17f9b4bb5e399ed4439bf85f198ba1476473c98" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.766573 4992 scope.go:117] "RemoveContainer" containerID="16c3322b8796d7b6a6fba9e2d760ce5c71fba07e153ce98649e53d312d65c58c" Oct 04 05:03:10 crc kubenswrapper[4992]: I1004 05:03:10.799677 4992 scope.go:117] "RemoveContainer" containerID="33de08d820fa120fa3e4c06b5ec746bda80c31ac74728f3c9c351b9bf084588c" Oct 04 05:03:16 crc kubenswrapper[4992]: I1004 05:03:16.320707 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:03:16 crc kubenswrapper[4992]: E1004 05:03:16.321480 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:03:29 crc kubenswrapper[4992]: I1004 05:03:29.320060 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:03:29 crc kubenswrapper[4992]: E1004 05:03:29.321060 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:03:43 crc kubenswrapper[4992]: I1004 05:03:43.319185 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:03:43 crc kubenswrapper[4992]: E1004 05:03:43.319894 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:03:52 crc kubenswrapper[4992]: I1004 05:03:52.044337 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ln7d7"] Oct 04 05:03:52 crc kubenswrapper[4992]: I1004 05:03:52.052709 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ln7d7"] Oct 04 05:03:52 crc kubenswrapper[4992]: I1004 05:03:52.330128 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd8feb2f-147b-43d8-ab87-58703170e959" path="/var/lib/kubelet/pods/dd8feb2f-147b-43d8-ab87-58703170e959/volumes" Oct 04 05:03:54 crc kubenswrapper[4992]: I1004 05:03:54.818674 4992 generic.go:334] "Generic (PLEG): container finished" podID="886d41dd-15e8-48ce-bbe9-f8df87efabba" containerID="b51edca6b10bdd6085cec62bd22cbea873c2d07971a951ed90eda2c5f2545090" exitCode=0 Oct 04 05:03:54 crc kubenswrapper[4992]: I1004 05:03:54.818764 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" event={"ID":"886d41dd-15e8-48ce-bbe9-f8df87efabba","Type":"ContainerDied","Data":"b51edca6b10bdd6085cec62bd22cbea873c2d07971a951ed90eda2c5f2545090"} Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.195419 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.281397 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-inventory\") pod \"886d41dd-15e8-48ce-bbe9-f8df87efabba\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.281735 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-ssh-key\") pod \"886d41dd-15e8-48ce-bbe9-f8df87efabba\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.281822 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctwd8\" (UniqueName: \"kubernetes.io/projected/886d41dd-15e8-48ce-bbe9-f8df87efabba-kube-api-access-ctwd8\") pod \"886d41dd-15e8-48ce-bbe9-f8df87efabba\" (UID: \"886d41dd-15e8-48ce-bbe9-f8df87efabba\") " Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.286748 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/886d41dd-15e8-48ce-bbe9-f8df87efabba-kube-api-access-ctwd8" (OuterVolumeSpecName: "kube-api-access-ctwd8") pod "886d41dd-15e8-48ce-bbe9-f8df87efabba" (UID: "886d41dd-15e8-48ce-bbe9-f8df87efabba"). InnerVolumeSpecName "kube-api-access-ctwd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.307061 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-inventory" (OuterVolumeSpecName: "inventory") pod "886d41dd-15e8-48ce-bbe9-f8df87efabba" (UID: "886d41dd-15e8-48ce-bbe9-f8df87efabba"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.312327 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "886d41dd-15e8-48ce-bbe9-f8df87efabba" (UID: "886d41dd-15e8-48ce-bbe9-f8df87efabba"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.384775 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.384915 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/886d41dd-15e8-48ce-bbe9-f8df87efabba-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.385093 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctwd8\" (UniqueName: \"kubernetes.io/projected/886d41dd-15e8-48ce-bbe9-f8df87efabba-kube-api-access-ctwd8\") on node \"crc\" DevicePath \"\"" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.837760 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" event={"ID":"886d41dd-15e8-48ce-bbe9-f8df87efabba","Type":"ContainerDied","Data":"e4249e9dd8d558a676adeba7e13bea7d435208dd76258e13b31a6de7bd3e2d81"} Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.837789 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-cr45c" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.837802 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4249e9dd8d558a676adeba7e13bea7d435208dd76258e13b31a6de7bd3e2d81" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.920844 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v"] Oct 04 05:03:56 crc kubenswrapper[4992]: E1004 05:03:56.921334 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886d41dd-15e8-48ce-bbe9-f8df87efabba" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.921380 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="886d41dd-15e8-48ce-bbe9-f8df87efabba" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.921566 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="886d41dd-15e8-48ce-bbe9-f8df87efabba" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.922284 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.926661 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.926805 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.927047 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.927137 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.944325 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v"] Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.994952 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.995043 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:56 crc kubenswrapper[4992]: I1004 05:03:56.995270 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvvv2\" (UniqueName: \"kubernetes.io/projected/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-kube-api-access-fvvv2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.096885 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.097206 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.097279 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvvv2\" (UniqueName: \"kubernetes.io/projected/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-kube-api-access-fvvv2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.105530 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.109451 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.118674 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvvv2\" (UniqueName: \"kubernetes.io/projected/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-kube-api-access-fvvv2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-n554v\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.245244 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.322582 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:03:57 crc kubenswrapper[4992]: E1004 05:03:57.323136 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.777019 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v"] Oct 04 05:03:57 crc kubenswrapper[4992]: I1004 05:03:57.847206 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" event={"ID":"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe","Type":"ContainerStarted","Data":"659dc027fbe2699e47705bc64ff9fab379339e61dea814e5a171782b158293ef"} Oct 04 05:03:58 crc kubenswrapper[4992]: I1004 05:03:58.863756 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" event={"ID":"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe","Type":"ContainerStarted","Data":"24d1c30d4c5e68fae320e1ac52425d14eef9335913727ac14bf8ed5acca573a6"} Oct 04 05:03:58 crc kubenswrapper[4992]: I1004 05:03:58.884919 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" podStartSLOduration=2.286658204 podStartE2EDuration="2.88489682s" podCreationTimestamp="2025-10-04 05:03:56 +0000 UTC" firstStartedPulling="2025-10-04 05:03:57.785057977 +0000 UTC m=+1871.632733455" lastFinishedPulling="2025-10-04 05:03:58.383296593 +0000 UTC m=+1872.230972071" observedRunningTime="2025-10-04 05:03:58.878797928 +0000 UTC m=+1872.726473406" watchObservedRunningTime="2025-10-04 05:03:58.88489682 +0000 UTC m=+1872.732572308" Oct 04 05:04:02 crc kubenswrapper[4992]: I1004 05:04:02.030340 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-5578x"] Oct 04 05:04:02 crc kubenswrapper[4992]: I1004 05:04:02.037338 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-5578x"] Oct 04 05:04:02 crc kubenswrapper[4992]: I1004 05:04:02.330417 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2f4ea76-03f4-49e2-99d8-af57dabb6372" path="/var/lib/kubelet/pods/d2f4ea76-03f4-49e2-99d8-af57dabb6372/volumes" Oct 04 05:04:03 crc kubenswrapper[4992]: I1004 05:04:03.026507 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-srrcj"] Oct 04 05:04:03 crc kubenswrapper[4992]: I1004 05:04:03.038439 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-srrcj"] Oct 04 05:04:04 crc kubenswrapper[4992]: I1004 05:04:04.331739 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15d2bb9c-fea2-48e2-aac2-3f07c2813b25" path="/var/lib/kubelet/pods/15d2bb9c-fea2-48e2-aac2-3f07c2813b25/volumes" Oct 04 05:04:07 crc kubenswrapper[4992]: I1004 05:04:07.032103 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-zcz9v"] Oct 04 05:04:07 crc kubenswrapper[4992]: I1004 05:04:07.041391 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-zcz9v"] Oct 04 05:04:08 crc kubenswrapper[4992]: I1004 05:04:08.330133 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a96c84d-f51b-48c0-9fa8-6040ed4f40f6" path="/var/lib/kubelet/pods/0a96c84d-f51b-48c0-9fa8-6040ed4f40f6/volumes" Oct 04 05:04:09 crc kubenswrapper[4992]: I1004 05:04:09.318715 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:04:09 crc kubenswrapper[4992]: E1004 05:04:09.319381 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:04:11 crc kubenswrapper[4992]: I1004 05:04:11.073180 4992 scope.go:117] "RemoveContainer" containerID="debf3cba486bfdfb01da926716ae7e614f2f71ca37382b9cabbcb1fb44e12bb8" Oct 04 05:04:11 crc kubenswrapper[4992]: I1004 05:04:11.125106 4992 scope.go:117] "RemoveContainer" containerID="37d9da7f7b13e8017ebf021988b1d3a87fffa72d2f31f182996f8348347bd403" Oct 04 05:04:11 crc kubenswrapper[4992]: I1004 05:04:11.169378 4992 scope.go:117] "RemoveContainer" containerID="a708ae19fdcb6154dfbe928d522052f3b71af3f67a4382c3650b02f3c29e2a7c" Oct 04 05:04:11 crc kubenswrapper[4992]: I1004 05:04:11.206017 4992 scope.go:117] "RemoveContainer" containerID="00cff6f979663a239fcdc719ccbebebc14bbf88001e8f681e8cb7dfdab02ad74" Oct 04 05:04:15 crc kubenswrapper[4992]: I1004 05:04:15.040346 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-9pjhh"] Oct 04 05:04:15 crc kubenswrapper[4992]: I1004 05:04:15.047880 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-9pjhh"] Oct 04 05:04:16 crc kubenswrapper[4992]: I1004 05:04:16.329883 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb31ac75-c92d-4b64-87b0-b5dfd194617b" path="/var/lib/kubelet/pods/eb31ac75-c92d-4b64-87b0-b5dfd194617b/volumes" Oct 04 05:04:21 crc kubenswrapper[4992]: I1004 05:04:21.319822 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:04:21 crc kubenswrapper[4992]: E1004 05:04:21.320508 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:04:35 crc kubenswrapper[4992]: I1004 05:04:35.320033 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:04:35 crc kubenswrapper[4992]: E1004 05:04:35.321045 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.037403 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-5pxw8"] Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.050048 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-h7lnf"] Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.058609 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-62css"] Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.065950 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-5pxw8"] Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.075818 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-62css"] Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.086967 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-h7lnf"] Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.341351 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3343348-309c-4770-8459-977ee71a5571" path="/var/lib/kubelet/pods/c3343348-309c-4770-8459-977ee71a5571/volumes" Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.342987 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5340c92-9a3c-42f7-abbe-1f5e1cd2572b" path="/var/lib/kubelet/pods/c5340c92-9a3c-42f7-abbe-1f5e1cd2572b/volumes" Oct 04 05:04:42 crc kubenswrapper[4992]: I1004 05:04:42.345728 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3b4bd43-8554-475a-9cc6-af381cc236e4" path="/var/lib/kubelet/pods/e3b4bd43-8554-475a-9cc6-af381cc236e4/volumes" Oct 04 05:04:50 crc kubenswrapper[4992]: I1004 05:04:50.319763 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:04:50 crc kubenswrapper[4992]: E1004 05:04:50.320594 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:04:51 crc kubenswrapper[4992]: I1004 05:04:51.047759 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0d44-account-create-dr6kn"] Oct 04 05:04:51 crc kubenswrapper[4992]: I1004 05:04:51.055339 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-663d-account-create-vr8wb"] Oct 04 05:04:51 crc kubenswrapper[4992]: I1004 05:04:51.064907 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-663d-account-create-vr8wb"] Oct 04 05:04:51 crc kubenswrapper[4992]: I1004 05:04:51.071336 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0d44-account-create-dr6kn"] Oct 04 05:04:52 crc kubenswrapper[4992]: I1004 05:04:52.030485 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1181-account-create-7s4m2"] Oct 04 05:04:52 crc kubenswrapper[4992]: I1004 05:04:52.038936 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1181-account-create-7s4m2"] Oct 04 05:04:52 crc kubenswrapper[4992]: I1004 05:04:52.330402 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ab7a5be-31da-432b-9ceb-94b52df3932b" path="/var/lib/kubelet/pods/9ab7a5be-31da-432b-9ceb-94b52df3932b/volumes" Oct 04 05:04:52 crc kubenswrapper[4992]: I1004 05:04:52.331217 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0298f1a-b260-4298-8c05-ffc8ac1b2d08" path="/var/lib/kubelet/pods/c0298f1a-b260-4298-8c05-ffc8ac1b2d08/volumes" Oct 04 05:04:52 crc kubenswrapper[4992]: I1004 05:04:52.332023 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d75a804a-5ed9-4f5c-8426-85d49905f065" path="/var/lib/kubelet/pods/d75a804a-5ed9-4f5c-8426-85d49905f065/volumes" Oct 04 05:05:04 crc kubenswrapper[4992]: I1004 05:05:04.332511 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:05:04 crc kubenswrapper[4992]: E1004 05:05:04.335156 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:05:11 crc kubenswrapper[4992]: I1004 05:05:11.335039 4992 scope.go:117] "RemoveContainer" containerID="3a3061ba0c0c8aa15a6af5f3ff733de50ef6020a7a100ca1aa4a939fb027c7af" Oct 04 05:05:11 crc kubenswrapper[4992]: I1004 05:05:11.396068 4992 scope.go:117] "RemoveContainer" containerID="8cff7f33ca8498f10550ca9458b0415198e1e1b938303ee092bf4955ebc4c8b8" Oct 04 05:05:11 crc kubenswrapper[4992]: I1004 05:05:11.423378 4992 scope.go:117] "RemoveContainer" containerID="6d064c4627574d8dfe5a62ac0c60602fbca42898f81558091d71052804a329e4" Oct 04 05:05:11 crc kubenswrapper[4992]: I1004 05:05:11.461628 4992 scope.go:117] "RemoveContainer" containerID="d463686a8b7b1c4500d12116d647ebe0456ddb474581d901a3f5e2ea0ea04ccd" Oct 04 05:05:11 crc kubenswrapper[4992]: I1004 05:05:11.524956 4992 scope.go:117] "RemoveContainer" containerID="e83cd842770012a1ea1f2aa1079f4f2eadf00df7782980e6091c81a4da41160a" Oct 04 05:05:11 crc kubenswrapper[4992]: I1004 05:05:11.546317 4992 scope.go:117] "RemoveContainer" containerID="2019828bb9888d08b01787093f8c88dccc9f6fd7ebc29f21a6c644942b216844" Oct 04 05:05:11 crc kubenswrapper[4992]: I1004 05:05:11.592284 4992 scope.go:117] "RemoveContainer" containerID="e288880fc1464923ab57c5c368f1b090af95a43477c93f6c31a2e4b80f40ac42" Oct 04 05:05:15 crc kubenswrapper[4992]: I1004 05:05:15.319964 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:05:15 crc kubenswrapper[4992]: E1004 05:05:15.320388 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:05:15 crc kubenswrapper[4992]: I1004 05:05:15.617837 4992 generic.go:334] "Generic (PLEG): container finished" podID="d5fb9b36-df2c-4e2f-816c-a455cd91d3fe" containerID="24d1c30d4c5e68fae320e1ac52425d14eef9335913727ac14bf8ed5acca573a6" exitCode=0 Oct 04 05:05:15 crc kubenswrapper[4992]: I1004 05:05:15.617882 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" event={"ID":"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe","Type":"ContainerDied","Data":"24d1c30d4c5e68fae320e1ac52425d14eef9335913727ac14bf8ed5acca573a6"} Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.023500 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.120228 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-ssh-key\") pod \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.120305 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvvv2\" (UniqueName: \"kubernetes.io/projected/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-kube-api-access-fvvv2\") pod \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.120502 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-inventory\") pod \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\" (UID: \"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe\") " Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.128593 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-kube-api-access-fvvv2" (OuterVolumeSpecName: "kube-api-access-fvvv2") pod "d5fb9b36-df2c-4e2f-816c-a455cd91d3fe" (UID: "d5fb9b36-df2c-4e2f-816c-a455cd91d3fe"). InnerVolumeSpecName "kube-api-access-fvvv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.152499 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-inventory" (OuterVolumeSpecName: "inventory") pod "d5fb9b36-df2c-4e2f-816c-a455cd91d3fe" (UID: "d5fb9b36-df2c-4e2f-816c-a455cd91d3fe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.154955 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d5fb9b36-df2c-4e2f-816c-a455cd91d3fe" (UID: "d5fb9b36-df2c-4e2f-816c-a455cd91d3fe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.222619 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.222655 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.222666 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvvv2\" (UniqueName: \"kubernetes.io/projected/d5fb9b36-df2c-4e2f-816c-a455cd91d3fe-kube-api-access-fvvv2\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.637322 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" event={"ID":"d5fb9b36-df2c-4e2f-816c-a455cd91d3fe","Type":"ContainerDied","Data":"659dc027fbe2699e47705bc64ff9fab379339e61dea814e5a171782b158293ef"} Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.637425 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="659dc027fbe2699e47705bc64ff9fab379339e61dea814e5a171782b158293ef" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.637440 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-n554v" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.717384 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f"] Oct 04 05:05:17 crc kubenswrapper[4992]: E1004 05:05:17.718121 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5fb9b36-df2c-4e2f-816c-a455cd91d3fe" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.718146 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5fb9b36-df2c-4e2f-816c-a455cd91d3fe" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.718382 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5fb9b36-df2c-4e2f-816c-a455cd91d3fe" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.719009 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.720739 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.721076 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.725275 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.725484 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.731694 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f"] Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.832768 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.833022 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.833214 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dwpj\" (UniqueName: \"kubernetes.io/projected/9d3b0970-11af-4754-8ad0-beb5e97fd149-kube-api-access-2dwpj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.934573 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.934664 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dwpj\" (UniqueName: \"kubernetes.io/projected/9d3b0970-11af-4754-8ad0-beb5e97fd149-kube-api-access-2dwpj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.934725 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.941236 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.941702 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:17 crc kubenswrapper[4992]: I1004 05:05:17.949158 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dwpj\" (UniqueName: \"kubernetes.io/projected/9d3b0970-11af-4754-8ad0-beb5e97fd149-kube-api-access-2dwpj\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:18 crc kubenswrapper[4992]: I1004 05:05:18.035872 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:18 crc kubenswrapper[4992]: I1004 05:05:18.602789 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f"] Oct 04 05:05:18 crc kubenswrapper[4992]: I1004 05:05:18.653749 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" event={"ID":"9d3b0970-11af-4754-8ad0-beb5e97fd149","Type":"ContainerStarted","Data":"199b3c36aef368fc59734fd03653f23f591c4bf1795d5f755fcb32ad919441b6"} Oct 04 05:05:19 crc kubenswrapper[4992]: I1004 05:05:19.664684 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" event={"ID":"9d3b0970-11af-4754-8ad0-beb5e97fd149","Type":"ContainerStarted","Data":"4c7f117479f68126373a4d70e1dee1d4211bd08cff173b667d05d593f4fc575b"} Oct 04 05:05:19 crc kubenswrapper[4992]: I1004 05:05:19.686026 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" podStartSLOduration=2.187816034 podStartE2EDuration="2.6860033s" podCreationTimestamp="2025-10-04 05:05:17 +0000 UTC" firstStartedPulling="2025-10-04 05:05:18.611001845 +0000 UTC m=+1952.458677313" lastFinishedPulling="2025-10-04 05:05:19.109189111 +0000 UTC m=+1952.956864579" observedRunningTime="2025-10-04 05:05:19.679096696 +0000 UTC m=+1953.526772174" watchObservedRunningTime="2025-10-04 05:05:19.6860033 +0000 UTC m=+1953.533678768" Oct 04 05:05:23 crc kubenswrapper[4992]: I1004 05:05:23.046677 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ctg5r"] Oct 04 05:05:23 crc kubenswrapper[4992]: I1004 05:05:23.054987 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ctg5r"] Oct 04 05:05:24 crc kubenswrapper[4992]: I1004 05:05:24.334726 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dbd79c9-221a-4665-9046-9c8fb90cfc04" path="/var/lib/kubelet/pods/2dbd79c9-221a-4665-9046-9c8fb90cfc04/volumes" Oct 04 05:05:24 crc kubenswrapper[4992]: I1004 05:05:24.712875 4992 generic.go:334] "Generic (PLEG): container finished" podID="9d3b0970-11af-4754-8ad0-beb5e97fd149" containerID="4c7f117479f68126373a4d70e1dee1d4211bd08cff173b667d05d593f4fc575b" exitCode=0 Oct 04 05:05:24 crc kubenswrapper[4992]: I1004 05:05:24.712980 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" event={"ID":"9d3b0970-11af-4754-8ad0-beb5e97fd149","Type":"ContainerDied","Data":"4c7f117479f68126373a4d70e1dee1d4211bd08cff173b667d05d593f4fc575b"} Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.150859 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.304287 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-ssh-key\") pod \"9d3b0970-11af-4754-8ad0-beb5e97fd149\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.304566 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dwpj\" (UniqueName: \"kubernetes.io/projected/9d3b0970-11af-4754-8ad0-beb5e97fd149-kube-api-access-2dwpj\") pod \"9d3b0970-11af-4754-8ad0-beb5e97fd149\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.304760 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-inventory\") pod \"9d3b0970-11af-4754-8ad0-beb5e97fd149\" (UID: \"9d3b0970-11af-4754-8ad0-beb5e97fd149\") " Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.324675 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3b0970-11af-4754-8ad0-beb5e97fd149-kube-api-access-2dwpj" (OuterVolumeSpecName: "kube-api-access-2dwpj") pod "9d3b0970-11af-4754-8ad0-beb5e97fd149" (UID: "9d3b0970-11af-4754-8ad0-beb5e97fd149"). InnerVolumeSpecName "kube-api-access-2dwpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.340555 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-inventory" (OuterVolumeSpecName: "inventory") pod "9d3b0970-11af-4754-8ad0-beb5e97fd149" (UID: "9d3b0970-11af-4754-8ad0-beb5e97fd149"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.347635 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9d3b0970-11af-4754-8ad0-beb5e97fd149" (UID: "9d3b0970-11af-4754-8ad0-beb5e97fd149"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.407394 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.407543 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9d3b0970-11af-4754-8ad0-beb5e97fd149-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.407700 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dwpj\" (UniqueName: \"kubernetes.io/projected/9d3b0970-11af-4754-8ad0-beb5e97fd149-kube-api-access-2dwpj\") on node \"crc\" DevicePath \"\"" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.733842 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" event={"ID":"9d3b0970-11af-4754-8ad0-beb5e97fd149","Type":"ContainerDied","Data":"199b3c36aef368fc59734fd03653f23f591c4bf1795d5f755fcb32ad919441b6"} Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.733893 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="199b3c36aef368fc59734fd03653f23f591c4bf1795d5f755fcb32ad919441b6" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.733982 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.806263 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5"] Oct 04 05:05:26 crc kubenswrapper[4992]: E1004 05:05:26.806663 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3b0970-11af-4754-8ad0-beb5e97fd149" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.806681 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3b0970-11af-4754-8ad0-beb5e97fd149" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.806851 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3b0970-11af-4754-8ad0-beb5e97fd149" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.807500 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.809673 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.809773 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.829352 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5"] Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.813302 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.813606 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.916677 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwjxs\" (UniqueName: \"kubernetes.io/projected/9af8d0eb-c228-489d-a923-ea44104df2dd-kube-api-access-nwjxs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.916747 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:26 crc kubenswrapper[4992]: I1004 05:05:26.916897 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.019015 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwjxs\" (UniqueName: \"kubernetes.io/projected/9af8d0eb-c228-489d-a923-ea44104df2dd-kube-api-access-nwjxs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.019114 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.019186 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.022942 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.023094 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.039840 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwjxs\" (UniqueName: \"kubernetes.io/projected/9af8d0eb-c228-489d-a923-ea44104df2dd-kube-api-access-nwjxs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-xlrz5\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.124804 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.666719 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5"] Oct 04 05:05:27 crc kubenswrapper[4992]: W1004 05:05:27.675508 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9af8d0eb_c228_489d_a923_ea44104df2dd.slice/crio-20693fe1a7fde0ce8ee0d9aba3f45e61a6e78e0de7c670f77de33a9cea0d4575 WatchSource:0}: Error finding container 20693fe1a7fde0ce8ee0d9aba3f45e61a6e78e0de7c670f77de33a9cea0d4575: Status 404 returned error can't find the container with id 20693fe1a7fde0ce8ee0d9aba3f45e61a6e78e0de7c670f77de33a9cea0d4575 Oct 04 05:05:27 crc kubenswrapper[4992]: I1004 05:05:27.749551 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" event={"ID":"9af8d0eb-c228-489d-a923-ea44104df2dd","Type":"ContainerStarted","Data":"20693fe1a7fde0ce8ee0d9aba3f45e61a6e78e0de7c670f77de33a9cea0d4575"} Oct 04 05:05:28 crc kubenswrapper[4992]: I1004 05:05:28.319200 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:05:28 crc kubenswrapper[4992]: E1004 05:05:28.320018 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:05:28 crc kubenswrapper[4992]: I1004 05:05:28.770457 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" event={"ID":"9af8d0eb-c228-489d-a923-ea44104df2dd","Type":"ContainerStarted","Data":"dbd6e69bf52c3002461564088ead68e076983c89a80d77664ee5813bc522da48"} Oct 04 05:05:28 crc kubenswrapper[4992]: I1004 05:05:28.786477 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" podStartSLOduration=2.108265116 podStartE2EDuration="2.786462165s" podCreationTimestamp="2025-10-04 05:05:26 +0000 UTC" firstStartedPulling="2025-10-04 05:05:27.67804798 +0000 UTC m=+1961.525723438" lastFinishedPulling="2025-10-04 05:05:28.356245019 +0000 UTC m=+1962.203920487" observedRunningTime="2025-10-04 05:05:28.784103952 +0000 UTC m=+1962.631779420" watchObservedRunningTime="2025-10-04 05:05:28.786462165 +0000 UTC m=+1962.634137633" Oct 04 05:05:40 crc kubenswrapper[4992]: I1004 05:05:40.319159 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:05:40 crc kubenswrapper[4992]: E1004 05:05:40.319808 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:05:46 crc kubenswrapper[4992]: I1004 05:05:46.045528 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-b5b2w"] Oct 04 05:05:46 crc kubenswrapper[4992]: I1004 05:05:46.055902 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-b5b2w"] Oct 04 05:05:46 crc kubenswrapper[4992]: I1004 05:05:46.328849 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30b06dc6-6c8d-407b-bd1f-96f05c45660b" path="/var/lib/kubelet/pods/30b06dc6-6c8d-407b-bd1f-96f05c45660b/volumes" Oct 04 05:05:55 crc kubenswrapper[4992]: I1004 05:05:55.320051 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:05:55 crc kubenswrapper[4992]: E1004 05:05:55.320761 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:06:03 crc kubenswrapper[4992]: I1004 05:06:03.037419 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ts9f9"] Oct 04 05:06:03 crc kubenswrapper[4992]: I1004 05:06:03.044911 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-ts9f9"] Oct 04 05:06:04 crc kubenswrapper[4992]: I1004 05:06:04.329888 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cc135c0-8c87-4802-836a-8c59def0eb0d" path="/var/lib/kubelet/pods/2cc135c0-8c87-4802-836a-8c59def0eb0d/volumes" Oct 04 05:06:05 crc kubenswrapper[4992]: I1004 05:06:05.083896 4992 generic.go:334] "Generic (PLEG): container finished" podID="9af8d0eb-c228-489d-a923-ea44104df2dd" containerID="dbd6e69bf52c3002461564088ead68e076983c89a80d77664ee5813bc522da48" exitCode=0 Oct 04 05:06:05 crc kubenswrapper[4992]: I1004 05:06:05.083939 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" event={"ID":"9af8d0eb-c228-489d-a923-ea44104df2dd","Type":"ContainerDied","Data":"dbd6e69bf52c3002461564088ead68e076983c89a80d77664ee5813bc522da48"} Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.516750 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.572495 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwjxs\" (UniqueName: \"kubernetes.io/projected/9af8d0eb-c228-489d-a923-ea44104df2dd-kube-api-access-nwjxs\") pod \"9af8d0eb-c228-489d-a923-ea44104df2dd\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.572629 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-inventory\") pod \"9af8d0eb-c228-489d-a923-ea44104df2dd\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.572768 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-ssh-key\") pod \"9af8d0eb-c228-489d-a923-ea44104df2dd\" (UID: \"9af8d0eb-c228-489d-a923-ea44104df2dd\") " Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.579068 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9af8d0eb-c228-489d-a923-ea44104df2dd-kube-api-access-nwjxs" (OuterVolumeSpecName: "kube-api-access-nwjxs") pod "9af8d0eb-c228-489d-a923-ea44104df2dd" (UID: "9af8d0eb-c228-489d-a923-ea44104df2dd"). InnerVolumeSpecName "kube-api-access-nwjxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.614767 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-inventory" (OuterVolumeSpecName: "inventory") pod "9af8d0eb-c228-489d-a923-ea44104df2dd" (UID: "9af8d0eb-c228-489d-a923-ea44104df2dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.616472 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9af8d0eb-c228-489d-a923-ea44104df2dd" (UID: "9af8d0eb-c228-489d-a923-ea44104df2dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.676046 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.676103 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9af8d0eb-c228-489d-a923-ea44104df2dd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:06 crc kubenswrapper[4992]: I1004 05:06:06.676126 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwjxs\" (UniqueName: \"kubernetes.io/projected/9af8d0eb-c228-489d-a923-ea44104df2dd-kube-api-access-nwjxs\") on node \"crc\" DevicePath \"\"" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.099963 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" event={"ID":"9af8d0eb-c228-489d-a923-ea44104df2dd","Type":"ContainerDied","Data":"20693fe1a7fde0ce8ee0d9aba3f45e61a6e78e0de7c670f77de33a9cea0d4575"} Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.100178 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20693fe1a7fde0ce8ee0d9aba3f45e61a6e78e0de7c670f77de33a9cea0d4575" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.100002 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-xlrz5" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.194849 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f"] Oct 04 05:06:07 crc kubenswrapper[4992]: E1004 05:06:07.195343 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af8d0eb-c228-489d-a923-ea44104df2dd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.195394 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af8d0eb-c228-489d-a923-ea44104df2dd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.195760 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af8d0eb-c228-489d-a923-ea44104df2dd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.196718 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.198867 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.198919 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.199181 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.199583 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.204958 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f"] Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.286369 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/72135fbe-8e57-4d94-98f5-c9677f852e76-kube-api-access-m8gfw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.286578 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.286628 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.319818 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:06:07 crc kubenswrapper[4992]: E1004 05:06:07.320200 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.388132 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.388192 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.388217 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/72135fbe-8e57-4d94-98f5-c9677f852e76-kube-api-access-m8gfw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.393154 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.395387 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.415970 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/72135fbe-8e57-4d94-98f5-c9677f852e76-kube-api-access-m8gfw\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:07 crc kubenswrapper[4992]: I1004 05:06:07.513219 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:06:08 crc kubenswrapper[4992]: I1004 05:06:08.204087 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f"] Oct 04 05:06:09 crc kubenswrapper[4992]: I1004 05:06:09.117926 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" event={"ID":"72135fbe-8e57-4d94-98f5-c9677f852e76","Type":"ContainerStarted","Data":"252db8cf2011c5d8b8c4e604a1406541a7d96ee6fffe400e3fe54b5f6cdb7ada"} Oct 04 05:06:09 crc kubenswrapper[4992]: I1004 05:06:09.118310 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" event={"ID":"72135fbe-8e57-4d94-98f5-c9677f852e76","Type":"ContainerStarted","Data":"26e4a27b9c06cbb41f060287df55c455848162694290245d9554d71fe16812ae"} Oct 04 05:06:09 crc kubenswrapper[4992]: I1004 05:06:09.140775 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" podStartSLOduration=1.593188668 podStartE2EDuration="2.140752989s" podCreationTimestamp="2025-10-04 05:06:07 +0000 UTC" firstStartedPulling="2025-10-04 05:06:08.213971551 +0000 UTC m=+2002.061647009" lastFinishedPulling="2025-10-04 05:06:08.761535862 +0000 UTC m=+2002.609211330" observedRunningTime="2025-10-04 05:06:09.130432755 +0000 UTC m=+2002.978108233" watchObservedRunningTime="2025-10-04 05:06:09.140752989 +0000 UTC m=+2002.988428477" Oct 04 05:06:11 crc kubenswrapper[4992]: I1004 05:06:11.730827 4992 scope.go:117] "RemoveContainer" containerID="ebce72ef6dca56815743ca8f5b44109b59750eef660a14e596c39c5216e461ae" Oct 04 05:06:11 crc kubenswrapper[4992]: I1004 05:06:11.796556 4992 scope.go:117] "RemoveContainer" containerID="1ceb7c240fb308cad875d20402abe1d34f87bdc00eb054ac85c8762e2ae05835" Oct 04 05:06:11 crc kubenswrapper[4992]: I1004 05:06:11.837831 4992 scope.go:117] "RemoveContainer" containerID="c5797424d349d2b1dbc6ee1f637579fd4ab0bd96477674c205fea64cb5bd6d88" Oct 04 05:06:18 crc kubenswrapper[4992]: I1004 05:06:18.319876 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:06:18 crc kubenswrapper[4992]: E1004 05:06:18.320720 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:06:30 crc kubenswrapper[4992]: I1004 05:06:30.319404 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:06:31 crc kubenswrapper[4992]: I1004 05:06:31.295963 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"3bca95cda3843b59172df471f72becf1de18ceb48613d521c6b0ccd2ca05ae02"} Oct 04 05:06:32 crc kubenswrapper[4992]: I1004 05:06:32.042902 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-cwhkv"] Oct 04 05:06:32 crc kubenswrapper[4992]: I1004 05:06:32.051497 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-cwhkv"] Oct 04 05:06:32 crc kubenswrapper[4992]: I1004 05:06:32.330064 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d" path="/var/lib/kubelet/pods/1b1f2d8a-1590-41e5-95ac-69ac4bc5dd7d/volumes" Oct 04 05:07:04 crc kubenswrapper[4992]: I1004 05:07:04.579074 4992 generic.go:334] "Generic (PLEG): container finished" podID="72135fbe-8e57-4d94-98f5-c9677f852e76" containerID="252db8cf2011c5d8b8c4e604a1406541a7d96ee6fffe400e3fe54b5f6cdb7ada" exitCode=2 Oct 04 05:07:04 crc kubenswrapper[4992]: I1004 05:07:04.579177 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" event={"ID":"72135fbe-8e57-4d94-98f5-c9677f852e76","Type":"ContainerDied","Data":"252db8cf2011c5d8b8c4e604a1406541a7d96ee6fffe400e3fe54b5f6cdb7ada"} Oct 04 05:07:05 crc kubenswrapper[4992]: I1004 05:07:05.974138 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.029297 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-ssh-key\") pod \"72135fbe-8e57-4d94-98f5-c9677f852e76\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.029803 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/72135fbe-8e57-4d94-98f5-c9677f852e76-kube-api-access-m8gfw\") pod \"72135fbe-8e57-4d94-98f5-c9677f852e76\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.029899 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-inventory\") pod \"72135fbe-8e57-4d94-98f5-c9677f852e76\" (UID: \"72135fbe-8e57-4d94-98f5-c9677f852e76\") " Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.035670 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72135fbe-8e57-4d94-98f5-c9677f852e76-kube-api-access-m8gfw" (OuterVolumeSpecName: "kube-api-access-m8gfw") pod "72135fbe-8e57-4d94-98f5-c9677f852e76" (UID: "72135fbe-8e57-4d94-98f5-c9677f852e76"). InnerVolumeSpecName "kube-api-access-m8gfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.056983 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-inventory" (OuterVolumeSpecName: "inventory") pod "72135fbe-8e57-4d94-98f5-c9677f852e76" (UID: "72135fbe-8e57-4d94-98f5-c9677f852e76"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.062097 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "72135fbe-8e57-4d94-98f5-c9677f852e76" (UID: "72135fbe-8e57-4d94-98f5-c9677f852e76"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.132877 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8gfw\" (UniqueName: \"kubernetes.io/projected/72135fbe-8e57-4d94-98f5-c9677f852e76-kube-api-access-m8gfw\") on node \"crc\" DevicePath \"\"" Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.133116 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.133185 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72135fbe-8e57-4d94-98f5-c9677f852e76-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.599651 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" event={"ID":"72135fbe-8e57-4d94-98f5-c9677f852e76","Type":"ContainerDied","Data":"26e4a27b9c06cbb41f060287df55c455848162694290245d9554d71fe16812ae"} Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.599686 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f" Oct 04 05:07:06 crc kubenswrapper[4992]: I1004 05:07:06.599701 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26e4a27b9c06cbb41f060287df55c455848162694290245d9554d71fe16812ae" Oct 04 05:07:11 crc kubenswrapper[4992]: I1004 05:07:11.968120 4992 scope.go:117] "RemoveContainer" containerID="ed5b2ca893c80e74fd5b3ee930d3a75e23f161b9e1839ef31cbce3468343f7de" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.031570 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm"] Oct 04 05:07:14 crc kubenswrapper[4992]: E1004 05:07:14.032258 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72135fbe-8e57-4d94-98f5-c9677f852e76" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.032273 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="72135fbe-8e57-4d94-98f5-c9677f852e76" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.032477 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="72135fbe-8e57-4d94-98f5-c9677f852e76" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.033139 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.035878 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.036276 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.038281 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.038485 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.045144 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm"] Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.203994 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.204302 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.204374 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb7zz\" (UniqueName: \"kubernetes.io/projected/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-kube-api-access-mb7zz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.307265 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.307306 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb7zz\" (UniqueName: \"kubernetes.io/projected/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-kube-api-access-mb7zz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.307422 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.313503 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.316852 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.324404 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb7zz\" (UniqueName: \"kubernetes.io/projected/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-kube-api-access-mb7zz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.359988 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:07:14 crc kubenswrapper[4992]: I1004 05:07:14.886694 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm"] Oct 04 05:07:15 crc kubenswrapper[4992]: I1004 05:07:15.677413 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" event={"ID":"5d87911a-e1f0-4c6c-a034-0b9e87faef4a","Type":"ContainerStarted","Data":"eba67042f38795cf7cc1f1338ffd85e1b0ad2e4bb4849d34247e64bd08c4f471"} Oct 04 05:07:16 crc kubenswrapper[4992]: I1004 05:07:16.690348 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" event={"ID":"5d87911a-e1f0-4c6c-a034-0b9e87faef4a","Type":"ContainerStarted","Data":"10560ec06f5c5561effc736ee95d271fce39acf3632d8e75f20a27fbc79a8bf2"} Oct 04 05:07:16 crc kubenswrapper[4992]: I1004 05:07:16.725015 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" podStartSLOduration=1.858355873 podStartE2EDuration="2.724993319s" podCreationTimestamp="2025-10-04 05:07:14 +0000 UTC" firstStartedPulling="2025-10-04 05:07:14.89424595 +0000 UTC m=+2068.741921418" lastFinishedPulling="2025-10-04 05:07:15.760883376 +0000 UTC m=+2069.608558864" observedRunningTime="2025-10-04 05:07:16.715329761 +0000 UTC m=+2070.563005259" watchObservedRunningTime="2025-10-04 05:07:16.724993319 +0000 UTC m=+2070.572668797" Oct 04 05:08:01 crc kubenswrapper[4992]: I1004 05:08:01.060984 4992 generic.go:334] "Generic (PLEG): container finished" podID="5d87911a-e1f0-4c6c-a034-0b9e87faef4a" containerID="10560ec06f5c5561effc736ee95d271fce39acf3632d8e75f20a27fbc79a8bf2" exitCode=0 Oct 04 05:08:01 crc kubenswrapper[4992]: I1004 05:08:01.061515 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" event={"ID":"5d87911a-e1f0-4c6c-a034-0b9e87faef4a","Type":"ContainerDied","Data":"10560ec06f5c5561effc736ee95d271fce39acf3632d8e75f20a27fbc79a8bf2"} Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.473910 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.628537 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-inventory\") pod \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.628653 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb7zz\" (UniqueName: \"kubernetes.io/projected/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-kube-api-access-mb7zz\") pod \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.628932 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-ssh-key\") pod \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\" (UID: \"5d87911a-e1f0-4c6c-a034-0b9e87faef4a\") " Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.635642 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-kube-api-access-mb7zz" (OuterVolumeSpecName: "kube-api-access-mb7zz") pod "5d87911a-e1f0-4c6c-a034-0b9e87faef4a" (UID: "5d87911a-e1f0-4c6c-a034-0b9e87faef4a"). InnerVolumeSpecName "kube-api-access-mb7zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.660228 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d87911a-e1f0-4c6c-a034-0b9e87faef4a" (UID: "5d87911a-e1f0-4c6c-a034-0b9e87faef4a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.660621 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-inventory" (OuterVolumeSpecName: "inventory") pod "5d87911a-e1f0-4c6c-a034-0b9e87faef4a" (UID: "5d87911a-e1f0-4c6c-a034-0b9e87faef4a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.730791 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.731031 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:02 crc kubenswrapper[4992]: I1004 05:08:02.731095 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb7zz\" (UniqueName: \"kubernetes.io/projected/5d87911a-e1f0-4c6c-a034-0b9e87faef4a-kube-api-access-mb7zz\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.079154 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" event={"ID":"5d87911a-e1f0-4c6c-a034-0b9e87faef4a","Type":"ContainerDied","Data":"eba67042f38795cf7cc1f1338ffd85e1b0ad2e4bb4849d34247e64bd08c4f471"} Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.079748 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eba67042f38795cf7cc1f1338ffd85e1b0ad2e4bb4849d34247e64bd08c4f471" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.079894 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.171256 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-96vsc"] Oct 04 05:08:03 crc kubenswrapper[4992]: E1004 05:08:03.172155 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d87911a-e1f0-4c6c-a034-0b9e87faef4a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.172184 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d87911a-e1f0-4c6c-a034-0b9e87faef4a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.176091 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d87911a-e1f0-4c6c-a034-0b9e87faef4a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.179760 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.184046 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.184138 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.184385 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.184771 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.206637 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-96vsc"] Oct 04 05:08:03 crc kubenswrapper[4992]: E1004 05:08:03.310217 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d87911a_e1f0_4c6c_a034_0b9e87faef4a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d87911a_e1f0_4c6c_a034_0b9e87faef4a.slice/crio-eba67042f38795cf7cc1f1338ffd85e1b0ad2e4bb4849d34247e64bd08c4f471\": RecentStats: unable to find data in memory cache]" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.342217 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.342783 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.342903 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97ltq\" (UniqueName: \"kubernetes.io/projected/04306ce6-821d-47e9-bfb1-765b7a067d44-kube-api-access-97ltq\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.444937 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.445025 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97ltq\" (UniqueName: \"kubernetes.io/projected/04306ce6-821d-47e9-bfb1-765b7a067d44-kube-api-access-97ltq\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.445063 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.450616 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.459142 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.462665 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97ltq\" (UniqueName: \"kubernetes.io/projected/04306ce6-821d-47e9-bfb1-765b7a067d44-kube-api-access-97ltq\") pod \"ssh-known-hosts-edpm-deployment-96vsc\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:03 crc kubenswrapper[4992]: I1004 05:08:03.513102 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:04 crc kubenswrapper[4992]: I1004 05:08:04.095782 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-96vsc"] Oct 04 05:08:04 crc kubenswrapper[4992]: I1004 05:08:04.102253 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:08:05 crc kubenswrapper[4992]: I1004 05:08:05.096173 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" event={"ID":"04306ce6-821d-47e9-bfb1-765b7a067d44","Type":"ContainerStarted","Data":"f50e0645abb65dcd155c9e72fd53e66ef3742773028e72976ab1e447a6d1649a"} Oct 04 05:08:06 crc kubenswrapper[4992]: I1004 05:08:06.107166 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" event={"ID":"04306ce6-821d-47e9-bfb1-765b7a067d44","Type":"ContainerStarted","Data":"4856ff6ce8115f7ba216c7c57af24c9dc4a82ee5901f52e05accc1a857c8bfda"} Oct 04 05:08:06 crc kubenswrapper[4992]: I1004 05:08:06.131695 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" podStartSLOduration=2.37286417 podStartE2EDuration="3.131674025s" podCreationTimestamp="2025-10-04 05:08:03 +0000 UTC" firstStartedPulling="2025-10-04 05:08:04.102062886 +0000 UTC m=+2117.949738354" lastFinishedPulling="2025-10-04 05:08:04.860872721 +0000 UTC m=+2118.708548209" observedRunningTime="2025-10-04 05:08:06.128781658 +0000 UTC m=+2119.976457156" watchObservedRunningTime="2025-10-04 05:08:06.131674025 +0000 UTC m=+2119.979349493" Oct 04 05:08:12 crc kubenswrapper[4992]: I1004 05:08:12.163236 4992 generic.go:334] "Generic (PLEG): container finished" podID="04306ce6-821d-47e9-bfb1-765b7a067d44" containerID="4856ff6ce8115f7ba216c7c57af24c9dc4a82ee5901f52e05accc1a857c8bfda" exitCode=0 Oct 04 05:08:12 crc kubenswrapper[4992]: I1004 05:08:12.163285 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" event={"ID":"04306ce6-821d-47e9-bfb1-765b7a067d44","Type":"ContainerDied","Data":"4856ff6ce8115f7ba216c7c57af24c9dc4a82ee5901f52e05accc1a857c8bfda"} Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.577875 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.754671 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-ssh-key-openstack-edpm-ipam\") pod \"04306ce6-821d-47e9-bfb1-765b7a067d44\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.754976 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-inventory-0\") pod \"04306ce6-821d-47e9-bfb1-765b7a067d44\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.755232 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97ltq\" (UniqueName: \"kubernetes.io/projected/04306ce6-821d-47e9-bfb1-765b7a067d44-kube-api-access-97ltq\") pod \"04306ce6-821d-47e9-bfb1-765b7a067d44\" (UID: \"04306ce6-821d-47e9-bfb1-765b7a067d44\") " Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.760732 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04306ce6-821d-47e9-bfb1-765b7a067d44-kube-api-access-97ltq" (OuterVolumeSpecName: "kube-api-access-97ltq") pod "04306ce6-821d-47e9-bfb1-765b7a067d44" (UID: "04306ce6-821d-47e9-bfb1-765b7a067d44"). InnerVolumeSpecName "kube-api-access-97ltq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.787089 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "04306ce6-821d-47e9-bfb1-765b7a067d44" (UID: "04306ce6-821d-47e9-bfb1-765b7a067d44"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.787663 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "04306ce6-821d-47e9-bfb1-765b7a067d44" (UID: "04306ce6-821d-47e9-bfb1-765b7a067d44"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.857266 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97ltq\" (UniqueName: \"kubernetes.io/projected/04306ce6-821d-47e9-bfb1-765b7a067d44-kube-api-access-97ltq\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.857295 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:13 crc kubenswrapper[4992]: I1004 05:08:13.857307 4992 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/04306ce6-821d-47e9-bfb1-765b7a067d44-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.181915 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" event={"ID":"04306ce6-821d-47e9-bfb1-765b7a067d44","Type":"ContainerDied","Data":"f50e0645abb65dcd155c9e72fd53e66ef3742773028e72976ab1e447a6d1649a"} Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.181957 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f50e0645abb65dcd155c9e72fd53e66ef3742773028e72976ab1e447a6d1649a" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.182002 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-96vsc" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.267400 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5"] Oct 04 05:08:14 crc kubenswrapper[4992]: E1004 05:08:14.267791 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04306ce6-821d-47e9-bfb1-765b7a067d44" containerName="ssh-known-hosts-edpm-deployment" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.269496 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="04306ce6-821d-47e9-bfb1-765b7a067d44" containerName="ssh-known-hosts-edpm-deployment" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.269826 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="04306ce6-821d-47e9-bfb1-765b7a067d44" containerName="ssh-known-hosts-edpm-deployment" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.270482 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.273540 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.273677 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.274176 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.275814 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.276222 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5"] Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.466353 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbqpl\" (UniqueName: \"kubernetes.io/projected/b57bbb97-a712-4ac2-841d-43b2053a009a-kube-api-access-dbqpl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.466481 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.466514 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.569065 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbqpl\" (UniqueName: \"kubernetes.io/projected/b57bbb97-a712-4ac2-841d-43b2053a009a-kube-api-access-dbqpl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.569199 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.569250 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.573384 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.574954 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.598147 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbqpl\" (UniqueName: \"kubernetes.io/projected/b57bbb97-a712-4ac2-841d-43b2053a009a-kube-api-access-dbqpl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-vtjm5\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:14 crc kubenswrapper[4992]: I1004 05:08:14.888438 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:15 crc kubenswrapper[4992]: I1004 05:08:15.398230 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5"] Oct 04 05:08:16 crc kubenswrapper[4992]: I1004 05:08:16.203969 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" event={"ID":"b57bbb97-a712-4ac2-841d-43b2053a009a","Type":"ContainerStarted","Data":"18dfccf0a9f1fffd1fd3d20d9e42bae0882c49bc633773f64ed5371d0e94a8b0"} Oct 04 05:08:17 crc kubenswrapper[4992]: I1004 05:08:17.216648 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" event={"ID":"b57bbb97-a712-4ac2-841d-43b2053a009a","Type":"ContainerStarted","Data":"813827baf569037bcf9021c2247990eb976fdba59d25cd320a8e26de6c5112cf"} Oct 04 05:08:17 crc kubenswrapper[4992]: I1004 05:08:17.237322 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" podStartSLOduration=2.374551567 podStartE2EDuration="3.23729717s" podCreationTimestamp="2025-10-04 05:08:14 +0000 UTC" firstStartedPulling="2025-10-04 05:08:15.403781105 +0000 UTC m=+2129.251456573" lastFinishedPulling="2025-10-04 05:08:16.266526708 +0000 UTC m=+2130.114202176" observedRunningTime="2025-10-04 05:08:17.232471101 +0000 UTC m=+2131.080146579" watchObservedRunningTime="2025-10-04 05:08:17.23729717 +0000 UTC m=+2131.084972638" Oct 04 05:08:25 crc kubenswrapper[4992]: I1004 05:08:25.285759 4992 generic.go:334] "Generic (PLEG): container finished" podID="b57bbb97-a712-4ac2-841d-43b2053a009a" containerID="813827baf569037bcf9021c2247990eb976fdba59d25cd320a8e26de6c5112cf" exitCode=0 Oct 04 05:08:25 crc kubenswrapper[4992]: I1004 05:08:25.285825 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" event={"ID":"b57bbb97-a712-4ac2-841d-43b2053a009a","Type":"ContainerDied","Data":"813827baf569037bcf9021c2247990eb976fdba59d25cd320a8e26de6c5112cf"} Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.701307 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.801204 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-inventory\") pod \"b57bbb97-a712-4ac2-841d-43b2053a009a\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.801288 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-ssh-key\") pod \"b57bbb97-a712-4ac2-841d-43b2053a009a\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.801884 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbqpl\" (UniqueName: \"kubernetes.io/projected/b57bbb97-a712-4ac2-841d-43b2053a009a-kube-api-access-dbqpl\") pod \"b57bbb97-a712-4ac2-841d-43b2053a009a\" (UID: \"b57bbb97-a712-4ac2-841d-43b2053a009a\") " Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.807841 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b57bbb97-a712-4ac2-841d-43b2053a009a-kube-api-access-dbqpl" (OuterVolumeSpecName: "kube-api-access-dbqpl") pod "b57bbb97-a712-4ac2-841d-43b2053a009a" (UID: "b57bbb97-a712-4ac2-841d-43b2053a009a"). InnerVolumeSpecName "kube-api-access-dbqpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.833453 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-inventory" (OuterVolumeSpecName: "inventory") pod "b57bbb97-a712-4ac2-841d-43b2053a009a" (UID: "b57bbb97-a712-4ac2-841d-43b2053a009a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.835643 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b57bbb97-a712-4ac2-841d-43b2053a009a" (UID: "b57bbb97-a712-4ac2-841d-43b2053a009a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.905552 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbqpl\" (UniqueName: \"kubernetes.io/projected/b57bbb97-a712-4ac2-841d-43b2053a009a-kube-api-access-dbqpl\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.905599 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:26 crc kubenswrapper[4992]: I1004 05:08:26.905611 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b57bbb97-a712-4ac2-841d-43b2053a009a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.308013 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" event={"ID":"b57bbb97-a712-4ac2-841d-43b2053a009a","Type":"ContainerDied","Data":"18dfccf0a9f1fffd1fd3d20d9e42bae0882c49bc633773f64ed5371d0e94a8b0"} Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.308468 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18dfccf0a9f1fffd1fd3d20d9e42bae0882c49bc633773f64ed5371d0e94a8b0" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.308111 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-vtjm5" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.454225 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb"] Oct 04 05:08:27 crc kubenswrapper[4992]: E1004 05:08:27.454762 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b57bbb97-a712-4ac2-841d-43b2053a009a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.454788 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b57bbb97-a712-4ac2-841d-43b2053a009a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.455014 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="b57bbb97-a712-4ac2-841d-43b2053a009a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.455901 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.458839 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.458944 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.459005 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.459057 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.470337 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb"] Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.619889 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.620132 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.620348 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lddvg\" (UniqueName: \"kubernetes.io/projected/ef40937c-226e-49ce-83eb-944481fe7a82-kube-api-access-lddvg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.722071 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.722134 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.722196 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lddvg\" (UniqueName: \"kubernetes.io/projected/ef40937c-226e-49ce-83eb-944481fe7a82-kube-api-access-lddvg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.728202 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.737109 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.745249 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lddvg\" (UniqueName: \"kubernetes.io/projected/ef40937c-226e-49ce-83eb-944481fe7a82-kube-api-access-lddvg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:27 crc kubenswrapper[4992]: I1004 05:08:27.778371 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:28 crc kubenswrapper[4992]: I1004 05:08:28.268597 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb"] Oct 04 05:08:28 crc kubenswrapper[4992]: W1004 05:08:28.273948 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef40937c_226e_49ce_83eb_944481fe7a82.slice/crio-ddfe4694208a6c24abac20da2b07038a0037a21ebc27de3487d5c4550fad2349 WatchSource:0}: Error finding container ddfe4694208a6c24abac20da2b07038a0037a21ebc27de3487d5c4550fad2349: Status 404 returned error can't find the container with id ddfe4694208a6c24abac20da2b07038a0037a21ebc27de3487d5c4550fad2349 Oct 04 05:08:28 crc kubenswrapper[4992]: I1004 05:08:28.330239 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" event={"ID":"ef40937c-226e-49ce-83eb-944481fe7a82","Type":"ContainerStarted","Data":"ddfe4694208a6c24abac20da2b07038a0037a21ebc27de3487d5c4550fad2349"} Oct 04 05:08:29 crc kubenswrapper[4992]: I1004 05:08:29.333643 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" event={"ID":"ef40937c-226e-49ce-83eb-944481fe7a82","Type":"ContainerStarted","Data":"652ebf2cfe667467f481df2606f9d6b05ad6f7d7fc1eb94bdfe9265b6436d5d1"} Oct 04 05:08:29 crc kubenswrapper[4992]: I1004 05:08:29.349985 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" podStartSLOduration=1.871421703 podStartE2EDuration="2.349957952s" podCreationTimestamp="2025-10-04 05:08:27 +0000 UTC" firstStartedPulling="2025-10-04 05:08:28.279905445 +0000 UTC m=+2142.127580913" lastFinishedPulling="2025-10-04 05:08:28.758441694 +0000 UTC m=+2142.606117162" observedRunningTime="2025-10-04 05:08:29.347812515 +0000 UTC m=+2143.195487983" watchObservedRunningTime="2025-10-04 05:08:29.349957952 +0000 UTC m=+2143.197633460" Oct 04 05:08:39 crc kubenswrapper[4992]: I1004 05:08:39.413773 4992 generic.go:334] "Generic (PLEG): container finished" podID="ef40937c-226e-49ce-83eb-944481fe7a82" containerID="652ebf2cfe667467f481df2606f9d6b05ad6f7d7fc1eb94bdfe9265b6436d5d1" exitCode=0 Oct 04 05:08:39 crc kubenswrapper[4992]: I1004 05:08:39.413852 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" event={"ID":"ef40937c-226e-49ce-83eb-944481fe7a82","Type":"ContainerDied","Data":"652ebf2cfe667467f481df2606f9d6b05ad6f7d7fc1eb94bdfe9265b6436d5d1"} Oct 04 05:08:40 crc kubenswrapper[4992]: I1004 05:08:40.810156 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:40 crc kubenswrapper[4992]: I1004 05:08:40.981077 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lddvg\" (UniqueName: \"kubernetes.io/projected/ef40937c-226e-49ce-83eb-944481fe7a82-kube-api-access-lddvg\") pod \"ef40937c-226e-49ce-83eb-944481fe7a82\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " Oct 04 05:08:40 crc kubenswrapper[4992]: I1004 05:08:40.981201 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-ssh-key\") pod \"ef40937c-226e-49ce-83eb-944481fe7a82\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " Oct 04 05:08:40 crc kubenswrapper[4992]: I1004 05:08:40.981445 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-inventory\") pod \"ef40937c-226e-49ce-83eb-944481fe7a82\" (UID: \"ef40937c-226e-49ce-83eb-944481fe7a82\") " Oct 04 05:08:40 crc kubenswrapper[4992]: I1004 05:08:40.986931 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef40937c-226e-49ce-83eb-944481fe7a82-kube-api-access-lddvg" (OuterVolumeSpecName: "kube-api-access-lddvg") pod "ef40937c-226e-49ce-83eb-944481fe7a82" (UID: "ef40937c-226e-49ce-83eb-944481fe7a82"). InnerVolumeSpecName "kube-api-access-lddvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.010872 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-inventory" (OuterVolumeSpecName: "inventory") pod "ef40937c-226e-49ce-83eb-944481fe7a82" (UID: "ef40937c-226e-49ce-83eb-944481fe7a82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.011177 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef40937c-226e-49ce-83eb-944481fe7a82" (UID: "ef40937c-226e-49ce-83eb-944481fe7a82"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.084216 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.084760 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lddvg\" (UniqueName: \"kubernetes.io/projected/ef40937c-226e-49ce-83eb-944481fe7a82-kube-api-access-lddvg\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.084863 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef40937c-226e-49ce-83eb-944481fe7a82-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.433458 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" event={"ID":"ef40937c-226e-49ce-83eb-944481fe7a82","Type":"ContainerDied","Data":"ddfe4694208a6c24abac20da2b07038a0037a21ebc27de3487d5c4550fad2349"} Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.433512 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddfe4694208a6c24abac20da2b07038a0037a21ebc27de3487d5c4550fad2349" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.433538 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.531105 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9"] Oct 04 05:08:41 crc kubenswrapper[4992]: E1004 05:08:41.532054 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef40937c-226e-49ce-83eb-944481fe7a82" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.532079 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef40937c-226e-49ce-83eb-944481fe7a82" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.532603 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef40937c-226e-49ce-83eb-944481fe7a82" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.533564 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.535283 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.535297 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.535850 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.536858 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.537167 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.537421 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.537710 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.537951 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.550580 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9"] Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.594931 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.594988 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595012 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595037 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595073 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595118 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595161 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvgtf\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-kube-api-access-qvgtf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595333 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595505 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595580 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595705 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595744 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595779 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.595820 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696722 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696781 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696804 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696837 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696876 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696902 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696938 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvgtf\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-kube-api-access-qvgtf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696961 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.696991 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.697016 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.697056 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.697075 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.697093 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.697112 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.702466 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.703020 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.703185 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.703504 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.704067 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.705048 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.705388 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.706102 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.706883 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.707054 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.707311 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.708196 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.716322 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvgtf\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-kube-api-access-qvgtf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.720130 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:41 crc kubenswrapper[4992]: I1004 05:08:41.852056 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:08:42 crc kubenswrapper[4992]: I1004 05:08:42.366416 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9"] Oct 04 05:08:42 crc kubenswrapper[4992]: I1004 05:08:42.443430 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" event={"ID":"3bafe4fd-0efb-4b1b-ba06-f5510110eb45","Type":"ContainerStarted","Data":"4eeb091f68ed0d6abe160448e8d5cd58c6b5c2d05380ab13f753f2a6ca646f2b"} Oct 04 05:08:43 crc kubenswrapper[4992]: I1004 05:08:43.455563 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" event={"ID":"3bafe4fd-0efb-4b1b-ba06-f5510110eb45","Type":"ContainerStarted","Data":"90058e482b44bf2b2c0ea961e4f77f78ec36d7461d1180d3f1a025cfc17e3bbe"} Oct 04 05:08:43 crc kubenswrapper[4992]: I1004 05:08:43.489040 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" podStartSLOduration=2.015863243 podStartE2EDuration="2.489023268s" podCreationTimestamp="2025-10-04 05:08:41 +0000 UTC" firstStartedPulling="2025-10-04 05:08:42.374663781 +0000 UTC m=+2156.222339249" lastFinishedPulling="2025-10-04 05:08:42.847823796 +0000 UTC m=+2156.695499274" observedRunningTime="2025-10-04 05:08:43.479771962 +0000 UTC m=+2157.327447450" watchObservedRunningTime="2025-10-04 05:08:43.489023268 +0000 UTC m=+2157.336698736" Oct 04 05:08:50 crc kubenswrapper[4992]: I1004 05:08:50.894650 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:08:50 crc kubenswrapper[4992]: I1004 05:08:50.895315 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.510762 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gvkxc"] Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.513457 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.519329 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gvkxc"] Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.570466 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-utilities\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.570605 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7grk\" (UniqueName: \"kubernetes.io/projected/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-kube-api-access-j7grk\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.570649 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-catalog-content\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.671950 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7grk\" (UniqueName: \"kubernetes.io/projected/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-kube-api-access-j7grk\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.672023 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-catalog-content\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.672156 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-utilities\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.672569 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-catalog-content\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.672680 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-utilities\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.702663 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7grk\" (UniqueName: \"kubernetes.io/projected/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-kube-api-access-j7grk\") pod \"redhat-operators-gvkxc\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.835413 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.894151 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:09:20 crc kubenswrapper[4992]: I1004 05:09:20.894212 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:09:21 crc kubenswrapper[4992]: I1004 05:09:21.353132 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gvkxc"] Oct 04 05:09:21 crc kubenswrapper[4992]: I1004 05:09:21.819752 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvkxc" event={"ID":"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f","Type":"ContainerStarted","Data":"909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443"} Oct 04 05:09:21 crc kubenswrapper[4992]: I1004 05:09:21.820143 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvkxc" event={"ID":"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f","Type":"ContainerStarted","Data":"a9a409b66ea1176f562d44a7e8364a804da0c17d00db01702f8aa0d6061536cd"} Oct 04 05:09:21 crc kubenswrapper[4992]: I1004 05:09:21.823107 4992 generic.go:334] "Generic (PLEG): container finished" podID="3bafe4fd-0efb-4b1b-ba06-f5510110eb45" containerID="90058e482b44bf2b2c0ea961e4f77f78ec36d7461d1180d3f1a025cfc17e3bbe" exitCode=0 Oct 04 05:09:21 crc kubenswrapper[4992]: I1004 05:09:21.823176 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" event={"ID":"3bafe4fd-0efb-4b1b-ba06-f5510110eb45","Type":"ContainerDied","Data":"90058e482b44bf2b2c0ea961e4f77f78ec36d7461d1180d3f1a025cfc17e3bbe"} Oct 04 05:09:22 crc kubenswrapper[4992]: I1004 05:09:22.832333 4992 generic.go:334] "Generic (PLEG): container finished" podID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerID="909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443" exitCode=0 Oct 04 05:09:22 crc kubenswrapper[4992]: I1004 05:09:22.832416 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvkxc" event={"ID":"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f","Type":"ContainerDied","Data":"909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443"} Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.261346 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.423668 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.423928 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-telemetry-combined-ca-bundle\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.423968 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.423999 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-ovn-default-certs-0\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424086 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-repo-setup-combined-ca-bundle\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424112 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-neutron-metadata-combined-ca-bundle\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424177 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424258 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-bootstrap-combined-ca-bundle\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424309 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ssh-key\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424353 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvgtf\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-kube-api-access-qvgtf\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424419 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-nova-combined-ca-bundle\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424444 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-inventory\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424471 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ovn-combined-ca-bundle\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.424494 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-libvirt-combined-ca-bundle\") pod \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\" (UID: \"3bafe4fd-0efb-4b1b-ba06-f5510110eb45\") " Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.431155 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.431677 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.431717 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.431774 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.431880 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.432684 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.433762 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.434403 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.434531 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-kube-api-access-qvgtf" (OuterVolumeSpecName: "kube-api-access-qvgtf") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "kube-api-access-qvgtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.434756 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.435466 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.435911 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.459423 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-inventory" (OuterVolumeSpecName: "inventory") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.464441 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3bafe4fd-0efb-4b1b-ba06-f5510110eb45" (UID: "3bafe4fd-0efb-4b1b-ba06-f5510110eb45"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526157 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526189 4992 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526202 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526220 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526230 4992 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526240 4992 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526249 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526260 4992 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526269 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526276 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvgtf\" (UniqueName: \"kubernetes.io/projected/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-kube-api-access-qvgtf\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526285 4992 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526293 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526301 4992 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.526310 4992 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bafe4fd-0efb-4b1b-ba06-f5510110eb45-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.849083 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvkxc" event={"ID":"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f","Type":"ContainerStarted","Data":"4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16"} Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.851729 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" event={"ID":"3bafe4fd-0efb-4b1b-ba06-f5510110eb45","Type":"ContainerDied","Data":"4eeb091f68ed0d6abe160448e8d5cd58c6b5c2d05380ab13f753f2a6ca646f2b"} Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.851868 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eeb091f68ed0d6abe160448e8d5cd58c6b5c2d05380ab13f753f2a6ca646f2b" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.851793 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.969276 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw"] Oct 04 05:09:23 crc kubenswrapper[4992]: E1004 05:09:23.969715 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bafe4fd-0efb-4b1b-ba06-f5510110eb45" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.969737 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bafe4fd-0efb-4b1b-ba06-f5510110eb45" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.970072 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bafe4fd-0efb-4b1b-ba06-f5510110eb45" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.975054 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.980872 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.980932 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.981000 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.981055 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.981099 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:09:23 crc kubenswrapper[4992]: I1004 05:09:23.984908 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw"] Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.136594 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.136706 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxwnr\" (UniqueName: \"kubernetes.io/projected/3185da64-244b-4f68-830f-f5d1cac08daa-kube-api-access-vxwnr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.136737 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.136805 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.136909 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3185da64-244b-4f68-830f-f5d1cac08daa-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.238281 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.238453 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3185da64-244b-4f68-830f-f5d1cac08daa-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.238516 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.238568 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxwnr\" (UniqueName: \"kubernetes.io/projected/3185da64-244b-4f68-830f-f5d1cac08daa-kube-api-access-vxwnr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.238592 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.240718 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3185da64-244b-4f68-830f-f5d1cac08daa-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.243664 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.243842 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.244201 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.263069 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxwnr\" (UniqueName: \"kubernetes.io/projected/3185da64-244b-4f68-830f-f5d1cac08daa-kube-api-access-vxwnr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-l6hlw\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.305680 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:09:24 crc kubenswrapper[4992]: W1004 05:09:24.831936 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3185da64_244b_4f68_830f_f5d1cac08daa.slice/crio-3353f0134e2c7e92e5673284e3e66aa3004d8bcfd8a4c623017651badcbb1eb1 WatchSource:0}: Error finding container 3353f0134e2c7e92e5673284e3e66aa3004d8bcfd8a4c623017651badcbb1eb1: Status 404 returned error can't find the container with id 3353f0134e2c7e92e5673284e3e66aa3004d8bcfd8a4c623017651badcbb1eb1 Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.840091 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw"] Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.859707 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" event={"ID":"3185da64-244b-4f68-830f-f5d1cac08daa","Type":"ContainerStarted","Data":"3353f0134e2c7e92e5673284e3e66aa3004d8bcfd8a4c623017651badcbb1eb1"} Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.861524 4992 generic.go:334] "Generic (PLEG): container finished" podID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerID="4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16" exitCode=0 Oct 04 05:09:24 crc kubenswrapper[4992]: I1004 05:09:24.861562 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvkxc" event={"ID":"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f","Type":"ContainerDied","Data":"4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16"} Oct 04 05:09:25 crc kubenswrapper[4992]: I1004 05:09:25.872916 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" event={"ID":"3185da64-244b-4f68-830f-f5d1cac08daa","Type":"ContainerStarted","Data":"9d9e426b8482d056241b80a645daa8f2aceaf93bedecd65821eaee5a6f9fbbfd"} Oct 04 05:09:25 crc kubenswrapper[4992]: I1004 05:09:25.875975 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvkxc" event={"ID":"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f","Type":"ContainerStarted","Data":"d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610"} Oct 04 05:09:25 crc kubenswrapper[4992]: I1004 05:09:25.901929 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" podStartSLOduration=2.480176154 podStartE2EDuration="2.90190986s" podCreationTimestamp="2025-10-04 05:09:23 +0000 UTC" firstStartedPulling="2025-10-04 05:09:24.835541361 +0000 UTC m=+2198.683216849" lastFinishedPulling="2025-10-04 05:09:25.257275087 +0000 UTC m=+2199.104950555" observedRunningTime="2025-10-04 05:09:25.894155573 +0000 UTC m=+2199.741831041" watchObservedRunningTime="2025-10-04 05:09:25.90190986 +0000 UTC m=+2199.749585328" Oct 04 05:09:25 crc kubenswrapper[4992]: I1004 05:09:25.922773 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gvkxc" podStartSLOduration=3.46910812 podStartE2EDuration="5.922757655s" podCreationTimestamp="2025-10-04 05:09:20 +0000 UTC" firstStartedPulling="2025-10-04 05:09:22.834904705 +0000 UTC m=+2196.682580173" lastFinishedPulling="2025-10-04 05:09:25.28855424 +0000 UTC m=+2199.136229708" observedRunningTime="2025-10-04 05:09:25.918357028 +0000 UTC m=+2199.766032496" watchObservedRunningTime="2025-10-04 05:09:25.922757655 +0000 UTC m=+2199.770433123" Oct 04 05:09:30 crc kubenswrapper[4992]: I1004 05:09:30.836195 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:30 crc kubenswrapper[4992]: I1004 05:09:30.836543 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:30 crc kubenswrapper[4992]: I1004 05:09:30.881632 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:30 crc kubenswrapper[4992]: I1004 05:09:30.956574 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:31 crc kubenswrapper[4992]: I1004 05:09:31.118444 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gvkxc"] Oct 04 05:09:32 crc kubenswrapper[4992]: I1004 05:09:32.931917 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gvkxc" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerName="registry-server" containerID="cri-o://d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610" gracePeriod=2 Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.340407 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.355216 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-utilities\") pod \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.355415 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7grk\" (UniqueName: \"kubernetes.io/projected/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-kube-api-access-j7grk\") pod \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.355622 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-catalog-content\") pod \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\" (UID: \"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f\") " Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.359031 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-utilities" (OuterVolumeSpecName: "utilities") pod "48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" (UID: "48ec742c-f0d3-4151-89d1-6d5b3f29cf6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.368493 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-kube-api-access-j7grk" (OuterVolumeSpecName: "kube-api-access-j7grk") pod "48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" (UID: "48ec742c-f0d3-4151-89d1-6d5b3f29cf6f"). InnerVolumeSpecName "kube-api-access-j7grk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.458148 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.458192 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7grk\" (UniqueName: \"kubernetes.io/projected/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-kube-api-access-j7grk\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.470501 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" (UID: "48ec742c-f0d3-4151-89d1-6d5b3f29cf6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.559587 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.944277 4992 generic.go:334] "Generic (PLEG): container finished" podID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerID="d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610" exitCode=0 Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.944328 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gvkxc" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.944348 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvkxc" event={"ID":"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f","Type":"ContainerDied","Data":"d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610"} Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.944599 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gvkxc" event={"ID":"48ec742c-f0d3-4151-89d1-6d5b3f29cf6f","Type":"ContainerDied","Data":"a9a409b66ea1176f562d44a7e8364a804da0c17d00db01702f8aa0d6061536cd"} Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.944623 4992 scope.go:117] "RemoveContainer" containerID="d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.965473 4992 scope.go:117] "RemoveContainer" containerID="4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16" Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.977045 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gvkxc"] Oct 04 05:09:33 crc kubenswrapper[4992]: I1004 05:09:33.983651 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gvkxc"] Oct 04 05:09:34 crc kubenswrapper[4992]: I1004 05:09:34.009761 4992 scope.go:117] "RemoveContainer" containerID="909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443" Oct 04 05:09:34 crc kubenswrapper[4992]: I1004 05:09:34.041240 4992 scope.go:117] "RemoveContainer" containerID="d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610" Oct 04 05:09:34 crc kubenswrapper[4992]: E1004 05:09:34.041794 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610\": container with ID starting with d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610 not found: ID does not exist" containerID="d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610" Oct 04 05:09:34 crc kubenswrapper[4992]: I1004 05:09:34.041848 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610"} err="failed to get container status \"d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610\": rpc error: code = NotFound desc = could not find container \"d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610\": container with ID starting with d3e3aceb6a16e37101b985302e14b066f4ef798665f684dfb05928d711c35610 not found: ID does not exist" Oct 04 05:09:34 crc kubenswrapper[4992]: I1004 05:09:34.041882 4992 scope.go:117] "RemoveContainer" containerID="4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16" Oct 04 05:09:34 crc kubenswrapper[4992]: E1004 05:09:34.042431 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16\": container with ID starting with 4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16 not found: ID does not exist" containerID="4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16" Oct 04 05:09:34 crc kubenswrapper[4992]: I1004 05:09:34.042462 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16"} err="failed to get container status \"4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16\": rpc error: code = NotFound desc = could not find container \"4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16\": container with ID starting with 4e416c3d0362ee14e1f0f4130b2b7f1433e1ecdd6f0174b2b0e755fa1caedc16 not found: ID does not exist" Oct 04 05:09:34 crc kubenswrapper[4992]: I1004 05:09:34.042486 4992 scope.go:117] "RemoveContainer" containerID="909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443" Oct 04 05:09:34 crc kubenswrapper[4992]: E1004 05:09:34.042924 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443\": container with ID starting with 909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443 not found: ID does not exist" containerID="909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443" Oct 04 05:09:34 crc kubenswrapper[4992]: I1004 05:09:34.042963 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443"} err="failed to get container status \"909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443\": rpc error: code = NotFound desc = could not find container \"909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443\": container with ID starting with 909ca92a3f8513ca94da82fc93ae68d7f3c836d381c194f882facbd781b44443 not found: ID does not exist" Oct 04 05:09:34 crc kubenswrapper[4992]: I1004 05:09:34.330747 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" path="/var/lib/kubelet/pods/48ec742c-f0d3-4151-89d1-6d5b3f29cf6f/volumes" Oct 04 05:09:50 crc kubenswrapper[4992]: I1004 05:09:50.894152 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:09:50 crc kubenswrapper[4992]: I1004 05:09:50.894744 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:09:50 crc kubenswrapper[4992]: I1004 05:09:50.894795 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 05:09:50 crc kubenswrapper[4992]: I1004 05:09:50.895657 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3bca95cda3843b59172df471f72becf1de18ceb48613d521c6b0ccd2ca05ae02"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:09:50 crc kubenswrapper[4992]: I1004 05:09:50.895723 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://3bca95cda3843b59172df471f72becf1de18ceb48613d521c6b0ccd2ca05ae02" gracePeriod=600 Oct 04 05:09:51 crc kubenswrapper[4992]: I1004 05:09:51.081308 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="3bca95cda3843b59172df471f72becf1de18ceb48613d521c6b0ccd2ca05ae02" exitCode=0 Oct 04 05:09:51 crc kubenswrapper[4992]: I1004 05:09:51.081409 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"3bca95cda3843b59172df471f72becf1de18ceb48613d521c6b0ccd2ca05ae02"} Oct 04 05:09:51 crc kubenswrapper[4992]: I1004 05:09:51.081529 4992 scope.go:117] "RemoveContainer" containerID="f4d564ba7d7ec98dc194e4982884f64e7438a67d36810723b3e2eedcf3a25817" Oct 04 05:09:52 crc kubenswrapper[4992]: I1004 05:09:52.099417 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505"} Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.226814 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5brcb"] Oct 04 05:09:53 crc kubenswrapper[4992]: E1004 05:09:53.227246 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerName="extract-content" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.227262 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerName="extract-content" Oct 04 05:09:53 crc kubenswrapper[4992]: E1004 05:09:53.227298 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerName="extract-utilities" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.227308 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerName="extract-utilities" Oct 04 05:09:53 crc kubenswrapper[4992]: E1004 05:09:53.227328 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerName="registry-server" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.227336 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerName="registry-server" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.227583 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="48ec742c-f0d3-4151-89d1-6d5b3f29cf6f" containerName="registry-server" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.234897 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.245341 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5brcb"] Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.325147 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-catalog-content\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.325546 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxh88\" (UniqueName: \"kubernetes.io/projected/ed4c429a-c948-40d0-98e7-9c057eb8ba24-kube-api-access-jxh88\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.325597 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-utilities\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.427217 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxh88\" (UniqueName: \"kubernetes.io/projected/ed4c429a-c948-40d0-98e7-9c057eb8ba24-kube-api-access-jxh88\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.427384 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-utilities\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.427625 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-catalog-content\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.428273 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-utilities\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.428379 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-catalog-content\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.448444 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxh88\" (UniqueName: \"kubernetes.io/projected/ed4c429a-c948-40d0-98e7-9c057eb8ba24-kube-api-access-jxh88\") pod \"certified-operators-5brcb\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:53 crc kubenswrapper[4992]: I1004 05:09:53.555513 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:09:54 crc kubenswrapper[4992]: I1004 05:09:54.116006 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5brcb"] Oct 04 05:09:54 crc kubenswrapper[4992]: I1004 05:09:54.129044 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5brcb" event={"ID":"ed4c429a-c948-40d0-98e7-9c057eb8ba24","Type":"ContainerStarted","Data":"c5ec248f206651c24b9efa950608124c88163e48a4d97b6370b77db61d8f1482"} Oct 04 05:09:55 crc kubenswrapper[4992]: I1004 05:09:55.142089 4992 generic.go:334] "Generic (PLEG): container finished" podID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerID="606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132" exitCode=0 Oct 04 05:09:55 crc kubenswrapper[4992]: I1004 05:09:55.142173 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5brcb" event={"ID":"ed4c429a-c948-40d0-98e7-9c057eb8ba24","Type":"ContainerDied","Data":"606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132"} Oct 04 05:09:56 crc kubenswrapper[4992]: I1004 05:09:56.152388 4992 generic.go:334] "Generic (PLEG): container finished" podID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerID="12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae" exitCode=0 Oct 04 05:09:56 crc kubenswrapper[4992]: I1004 05:09:56.153008 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5brcb" event={"ID":"ed4c429a-c948-40d0-98e7-9c057eb8ba24","Type":"ContainerDied","Data":"12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae"} Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.021529 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ztnm6"] Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.024962 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.030731 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ztnm6"] Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.111854 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-utilities\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.112388 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx96t\" (UniqueName: \"kubernetes.io/projected/8b290458-e1ea-4fda-be55-1db8f373cc8f-kube-api-access-kx96t\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.112576 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-catalog-content\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.213995 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-utilities\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.214070 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx96t\" (UniqueName: \"kubernetes.io/projected/8b290458-e1ea-4fda-be55-1db8f373cc8f-kube-api-access-kx96t\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.214120 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-catalog-content\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.214646 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-catalog-content\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.214872 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-utilities\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.237331 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx96t\" (UniqueName: \"kubernetes.io/projected/8b290458-e1ea-4fda-be55-1db8f373cc8f-kube-api-access-kx96t\") pod \"community-operators-ztnm6\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.346287 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:09:57 crc kubenswrapper[4992]: I1004 05:09:57.901328 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ztnm6"] Oct 04 05:09:57 crc kubenswrapper[4992]: W1004 05:09:57.906143 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b290458_e1ea_4fda_be55_1db8f373cc8f.slice/crio-a95bb1b1ec4991e10ca82b81b6191378e3ddc3bf9234d50b01ece0d06231f9d6 WatchSource:0}: Error finding container a95bb1b1ec4991e10ca82b81b6191378e3ddc3bf9234d50b01ece0d06231f9d6: Status 404 returned error can't find the container with id a95bb1b1ec4991e10ca82b81b6191378e3ddc3bf9234d50b01ece0d06231f9d6 Oct 04 05:09:58 crc kubenswrapper[4992]: I1004 05:09:58.171472 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5brcb" event={"ID":"ed4c429a-c948-40d0-98e7-9c057eb8ba24","Type":"ContainerStarted","Data":"b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4"} Oct 04 05:09:58 crc kubenswrapper[4992]: I1004 05:09:58.173766 4992 generic.go:334] "Generic (PLEG): container finished" podID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerID="6f6cd2fcb6d9602d858a5e2725f331e4e385a0a8a8cdc8417ae51514412b25c5" exitCode=0 Oct 04 05:09:58 crc kubenswrapper[4992]: I1004 05:09:58.173812 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztnm6" event={"ID":"8b290458-e1ea-4fda-be55-1db8f373cc8f","Type":"ContainerDied","Data":"6f6cd2fcb6d9602d858a5e2725f331e4e385a0a8a8cdc8417ae51514412b25c5"} Oct 04 05:09:58 crc kubenswrapper[4992]: I1004 05:09:58.173842 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztnm6" event={"ID":"8b290458-e1ea-4fda-be55-1db8f373cc8f","Type":"ContainerStarted","Data":"a95bb1b1ec4991e10ca82b81b6191378e3ddc3bf9234d50b01ece0d06231f9d6"} Oct 04 05:09:58 crc kubenswrapper[4992]: I1004 05:09:58.197207 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5brcb" podStartSLOduration=3.317311275 podStartE2EDuration="5.197185104s" podCreationTimestamp="2025-10-04 05:09:53 +0000 UTC" firstStartedPulling="2025-10-04 05:09:55.150058918 +0000 UTC m=+2228.997734386" lastFinishedPulling="2025-10-04 05:09:57.029932747 +0000 UTC m=+2230.877608215" observedRunningTime="2025-10-04 05:09:58.195677763 +0000 UTC m=+2232.043353241" watchObservedRunningTime="2025-10-04 05:09:58.197185104 +0000 UTC m=+2232.044860572" Oct 04 05:09:59 crc kubenswrapper[4992]: I1004 05:09:59.186215 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztnm6" event={"ID":"8b290458-e1ea-4fda-be55-1db8f373cc8f","Type":"ContainerStarted","Data":"a97006b34096e177e3b77ea78eaa3e729ca9beaf3a841f09d13b3b1553188921"} Oct 04 05:10:00 crc kubenswrapper[4992]: I1004 05:10:00.201307 4992 generic.go:334] "Generic (PLEG): container finished" podID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerID="a97006b34096e177e3b77ea78eaa3e729ca9beaf3a841f09d13b3b1553188921" exitCode=0 Oct 04 05:10:00 crc kubenswrapper[4992]: I1004 05:10:00.201439 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztnm6" event={"ID":"8b290458-e1ea-4fda-be55-1db8f373cc8f","Type":"ContainerDied","Data":"a97006b34096e177e3b77ea78eaa3e729ca9beaf3a841f09d13b3b1553188921"} Oct 04 05:10:01 crc kubenswrapper[4992]: I1004 05:10:01.216117 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztnm6" event={"ID":"8b290458-e1ea-4fda-be55-1db8f373cc8f","Type":"ContainerStarted","Data":"8b1569a1a9d71915ee550bece19c6c174925e72a1f695e955d9be0c51656b727"} Oct 04 05:10:03 crc kubenswrapper[4992]: I1004 05:10:03.555904 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:10:03 crc kubenswrapper[4992]: I1004 05:10:03.556850 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:10:03 crc kubenswrapper[4992]: I1004 05:10:03.610536 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:10:03 crc kubenswrapper[4992]: I1004 05:10:03.632059 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ztnm6" podStartSLOduration=5.146243384 podStartE2EDuration="7.63201467s" podCreationTimestamp="2025-10-04 05:09:56 +0000 UTC" firstStartedPulling="2025-10-04 05:09:58.175572697 +0000 UTC m=+2232.023248165" lastFinishedPulling="2025-10-04 05:10:00.661343983 +0000 UTC m=+2234.509019451" observedRunningTime="2025-10-04 05:10:01.240494147 +0000 UTC m=+2235.088169625" watchObservedRunningTime="2025-10-04 05:10:03.63201467 +0000 UTC m=+2237.479690148" Oct 04 05:10:04 crc kubenswrapper[4992]: I1004 05:10:04.290393 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:10:04 crc kubenswrapper[4992]: I1004 05:10:04.806061 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5brcb"] Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.260120 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5brcb" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerName="registry-server" containerID="cri-o://b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4" gracePeriod=2 Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.719011 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.730900 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxh88\" (UniqueName: \"kubernetes.io/projected/ed4c429a-c948-40d0-98e7-9c057eb8ba24-kube-api-access-jxh88\") pod \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.730993 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-catalog-content\") pod \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.731120 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-utilities\") pod \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\" (UID: \"ed4c429a-c948-40d0-98e7-9c057eb8ba24\") " Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.732131 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-utilities" (OuterVolumeSpecName: "utilities") pod "ed4c429a-c948-40d0-98e7-9c057eb8ba24" (UID: "ed4c429a-c948-40d0-98e7-9c057eb8ba24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.743663 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed4c429a-c948-40d0-98e7-9c057eb8ba24-kube-api-access-jxh88" (OuterVolumeSpecName: "kube-api-access-jxh88") pod "ed4c429a-c948-40d0-98e7-9c057eb8ba24" (UID: "ed4c429a-c948-40d0-98e7-9c057eb8ba24"). InnerVolumeSpecName "kube-api-access-jxh88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.806048 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed4c429a-c948-40d0-98e7-9c057eb8ba24" (UID: "ed4c429a-c948-40d0-98e7-9c057eb8ba24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.833823 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.833867 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed4c429a-c948-40d0-98e7-9c057eb8ba24-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:06 crc kubenswrapper[4992]: I1004 05:10:06.833887 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxh88\" (UniqueName: \"kubernetes.io/projected/ed4c429a-c948-40d0-98e7-9c057eb8ba24-kube-api-access-jxh88\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.270983 4992 generic.go:334] "Generic (PLEG): container finished" podID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerID="b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4" exitCode=0 Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.271120 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5brcb" event={"ID":"ed4c429a-c948-40d0-98e7-9c057eb8ba24","Type":"ContainerDied","Data":"b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4"} Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.271288 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5brcb" event={"ID":"ed4c429a-c948-40d0-98e7-9c057eb8ba24","Type":"ContainerDied","Data":"c5ec248f206651c24b9efa950608124c88163e48a4d97b6370b77db61d8f1482"} Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.271324 4992 scope.go:117] "RemoveContainer" containerID="b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.272494 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5brcb" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.294499 4992 scope.go:117] "RemoveContainer" containerID="12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.314174 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5brcb"] Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.324882 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5brcb"] Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.345323 4992 scope.go:117] "RemoveContainer" containerID="606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.346469 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.346534 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.370999 4992 scope.go:117] "RemoveContainer" containerID="b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4" Oct 04 05:10:07 crc kubenswrapper[4992]: E1004 05:10:07.371920 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4\": container with ID starting with b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4 not found: ID does not exist" containerID="b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.371954 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4"} err="failed to get container status \"b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4\": rpc error: code = NotFound desc = could not find container \"b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4\": container with ID starting with b5085355d9d40dbc55bce76a69e39d6b3fa3b116d655ab2d0adee8594eebb2b4 not found: ID does not exist" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.371979 4992 scope.go:117] "RemoveContainer" containerID="12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae" Oct 04 05:10:07 crc kubenswrapper[4992]: E1004 05:10:07.372432 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae\": container with ID starting with 12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae not found: ID does not exist" containerID="12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.372488 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae"} err="failed to get container status \"12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae\": rpc error: code = NotFound desc = could not find container \"12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae\": container with ID starting with 12388c84210c582062e39bf12c4e85ec6454cda8c4ba17fafd56c0b40332cfae not found: ID does not exist" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.372520 4992 scope.go:117] "RemoveContainer" containerID="606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132" Oct 04 05:10:07 crc kubenswrapper[4992]: E1004 05:10:07.372852 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132\": container with ID starting with 606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132 not found: ID does not exist" containerID="606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.372914 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132"} err="failed to get container status \"606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132\": rpc error: code = NotFound desc = could not find container \"606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132\": container with ID starting with 606b3db597484a66fea39b08f904a5016332ccebcc40992b151f469c95819132 not found: ID does not exist" Oct 04 05:10:07 crc kubenswrapper[4992]: I1004 05:10:07.400045 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:10:08 crc kubenswrapper[4992]: I1004 05:10:08.338096 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" path="/var/lib/kubelet/pods/ed4c429a-c948-40d0-98e7-9c057eb8ba24/volumes" Oct 04 05:10:08 crc kubenswrapper[4992]: I1004 05:10:08.339499 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:10:09 crc kubenswrapper[4992]: I1004 05:10:09.809738 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ztnm6"] Oct 04 05:10:10 crc kubenswrapper[4992]: I1004 05:10:10.300017 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ztnm6" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerName="registry-server" containerID="cri-o://8b1569a1a9d71915ee550bece19c6c174925e72a1f695e955d9be0c51656b727" gracePeriod=2 Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.313840 4992 generic.go:334] "Generic (PLEG): container finished" podID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerID="8b1569a1a9d71915ee550bece19c6c174925e72a1f695e955d9be0c51656b727" exitCode=0 Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.313880 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztnm6" event={"ID":"8b290458-e1ea-4fda-be55-1db8f373cc8f","Type":"ContainerDied","Data":"8b1569a1a9d71915ee550bece19c6c174925e72a1f695e955d9be0c51656b727"} Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.313906 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztnm6" event={"ID":"8b290458-e1ea-4fda-be55-1db8f373cc8f","Type":"ContainerDied","Data":"a95bb1b1ec4991e10ca82b81b6191378e3ddc3bf9234d50b01ece0d06231f9d6"} Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.313919 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a95bb1b1ec4991e10ca82b81b6191378e3ddc3bf9234d50b01ece0d06231f9d6" Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.383845 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.535022 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-utilities\") pod \"8b290458-e1ea-4fda-be55-1db8f373cc8f\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.535153 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-catalog-content\") pod \"8b290458-e1ea-4fda-be55-1db8f373cc8f\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.535234 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx96t\" (UniqueName: \"kubernetes.io/projected/8b290458-e1ea-4fda-be55-1db8f373cc8f-kube-api-access-kx96t\") pod \"8b290458-e1ea-4fda-be55-1db8f373cc8f\" (UID: \"8b290458-e1ea-4fda-be55-1db8f373cc8f\") " Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.536754 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-utilities" (OuterVolumeSpecName: "utilities") pod "8b290458-e1ea-4fda-be55-1db8f373cc8f" (UID: "8b290458-e1ea-4fda-be55-1db8f373cc8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.546657 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b290458-e1ea-4fda-be55-1db8f373cc8f-kube-api-access-kx96t" (OuterVolumeSpecName: "kube-api-access-kx96t") pod "8b290458-e1ea-4fda-be55-1db8f373cc8f" (UID: "8b290458-e1ea-4fda-be55-1db8f373cc8f"). InnerVolumeSpecName "kube-api-access-kx96t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.581014 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b290458-e1ea-4fda-be55-1db8f373cc8f" (UID: "8b290458-e1ea-4fda-be55-1db8f373cc8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.636634 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.636688 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b290458-e1ea-4fda-be55-1db8f373cc8f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:11 crc kubenswrapper[4992]: I1004 05:10:11.636701 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx96t\" (UniqueName: \"kubernetes.io/projected/8b290458-e1ea-4fda-be55-1db8f373cc8f-kube-api-access-kx96t\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:12 crc kubenswrapper[4992]: I1004 05:10:12.321768 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztnm6" Oct 04 05:10:12 crc kubenswrapper[4992]: I1004 05:10:12.363070 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ztnm6"] Oct 04 05:10:12 crc kubenswrapper[4992]: I1004 05:10:12.371201 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ztnm6"] Oct 04 05:10:14 crc kubenswrapper[4992]: I1004 05:10:14.335820 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" path="/var/lib/kubelet/pods/8b290458-e1ea-4fda-be55-1db8f373cc8f/volumes" Oct 04 05:10:26 crc kubenswrapper[4992]: I1004 05:10:26.436650 4992 generic.go:334] "Generic (PLEG): container finished" podID="3185da64-244b-4f68-830f-f5d1cac08daa" containerID="9d9e426b8482d056241b80a645daa8f2aceaf93bedecd65821eaee5a6f9fbbfd" exitCode=0 Oct 04 05:10:26 crc kubenswrapper[4992]: I1004 05:10:26.436756 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" event={"ID":"3185da64-244b-4f68-830f-f5d1cac08daa","Type":"ContainerDied","Data":"9d9e426b8482d056241b80a645daa8f2aceaf93bedecd65821eaee5a6f9fbbfd"} Oct 04 05:10:27 crc kubenswrapper[4992]: I1004 05:10:27.858494 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.039883 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory\") pod \"3185da64-244b-4f68-830f-f5d1cac08daa\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.039965 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxwnr\" (UniqueName: \"kubernetes.io/projected/3185da64-244b-4f68-830f-f5d1cac08daa-kube-api-access-vxwnr\") pod \"3185da64-244b-4f68-830f-f5d1cac08daa\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.040009 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ovn-combined-ca-bundle\") pod \"3185da64-244b-4f68-830f-f5d1cac08daa\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.040067 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ssh-key\") pod \"3185da64-244b-4f68-830f-f5d1cac08daa\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.040813 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3185da64-244b-4f68-830f-f5d1cac08daa-ovncontroller-config-0\") pod \"3185da64-244b-4f68-830f-f5d1cac08daa\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.046772 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3185da64-244b-4f68-830f-f5d1cac08daa-kube-api-access-vxwnr" (OuterVolumeSpecName: "kube-api-access-vxwnr") pod "3185da64-244b-4f68-830f-f5d1cac08daa" (UID: "3185da64-244b-4f68-830f-f5d1cac08daa"). InnerVolumeSpecName "kube-api-access-vxwnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.046787 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3185da64-244b-4f68-830f-f5d1cac08daa" (UID: "3185da64-244b-4f68-830f-f5d1cac08daa"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.067721 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3185da64-244b-4f68-830f-f5d1cac08daa-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "3185da64-244b-4f68-830f-f5d1cac08daa" (UID: "3185da64-244b-4f68-830f-f5d1cac08daa"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:10:28 crc kubenswrapper[4992]: E1004 05:10:28.070658 4992 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory podName:3185da64-244b-4f68-830f-f5d1cac08daa nodeName:}" failed. No retries permitted until 2025-10-04 05:10:28.570629153 +0000 UTC m=+2262.418304621 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory") pod "3185da64-244b-4f68-830f-f5d1cac08daa" (UID: "3185da64-244b-4f68-830f-f5d1cac08daa") : error deleting /var/lib/kubelet/pods/3185da64-244b-4f68-830f-f5d1cac08daa/volume-subpaths: remove /var/lib/kubelet/pods/3185da64-244b-4f68-830f-f5d1cac08daa/volume-subpaths: no such file or directory Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.074285 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3185da64-244b-4f68-830f-f5d1cac08daa" (UID: "3185da64-244b-4f68-830f-f5d1cac08daa"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.142842 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxwnr\" (UniqueName: \"kubernetes.io/projected/3185da64-244b-4f68-830f-f5d1cac08daa-kube-api-access-vxwnr\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.142881 4992 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.142894 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.142907 4992 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/3185da64-244b-4f68-830f-f5d1cac08daa-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.454109 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" event={"ID":"3185da64-244b-4f68-830f-f5d1cac08daa","Type":"ContainerDied","Data":"3353f0134e2c7e92e5673284e3e66aa3004d8bcfd8a4c623017651badcbb1eb1"} Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.454425 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3353f0134e2c7e92e5673284e3e66aa3004d8bcfd8a4c623017651badcbb1eb1" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.454252 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-l6hlw" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.583757 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69"] Oct 04 05:10:28 crc kubenswrapper[4992]: E1004 05:10:28.584396 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerName="extract-utilities" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.584430 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerName="extract-utilities" Oct 04 05:10:28 crc kubenswrapper[4992]: E1004 05:10:28.584450 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerName="extract-utilities" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.584465 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerName="extract-utilities" Oct 04 05:10:28 crc kubenswrapper[4992]: E1004 05:10:28.584491 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerName="registry-server" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.584503 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerName="registry-server" Oct 04 05:10:28 crc kubenswrapper[4992]: E1004 05:10:28.584529 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3185da64-244b-4f68-830f-f5d1cac08daa" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.584588 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3185da64-244b-4f68-830f-f5d1cac08daa" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 05:10:28 crc kubenswrapper[4992]: E1004 05:10:28.584622 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerName="registry-server" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.584634 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerName="registry-server" Oct 04 05:10:28 crc kubenswrapper[4992]: E1004 05:10:28.584666 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerName="extract-content" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.584677 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerName="extract-content" Oct 04 05:10:28 crc kubenswrapper[4992]: E1004 05:10:28.584708 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerName="extract-content" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.584720 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerName="extract-content" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.585034 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b290458-e1ea-4fda-be55-1db8f373cc8f" containerName="registry-server" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.585053 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed4c429a-c948-40d0-98e7-9c057eb8ba24" containerName="registry-server" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.585072 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3185da64-244b-4f68-830f-f5d1cac08daa" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.585951 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.588973 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.589556 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.597107 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69"] Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.652449 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory\") pod \"3185da64-244b-4f68-830f-f5d1cac08daa\" (UID: \"3185da64-244b-4f68-830f-f5d1cac08daa\") " Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.656055 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory" (OuterVolumeSpecName: "inventory") pod "3185da64-244b-4f68-830f-f5d1cac08daa" (UID: "3185da64-244b-4f68-830f-f5d1cac08daa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.754186 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.754252 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.754385 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k76c5\" (UniqueName: \"kubernetes.io/projected/241b4766-f3d7-4953-845a-6a898cea420e-kube-api-access-k76c5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.754452 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.754481 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.754511 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.754878 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3185da64-244b-4f68-830f-f5d1cac08daa-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.856266 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.856526 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.856663 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k76c5\" (UniqueName: \"kubernetes.io/projected/241b4766-f3d7-4953-845a-6a898cea420e-kube-api-access-k76c5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.856829 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.856945 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.857067 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.859862 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.860029 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.860265 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.860712 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.860948 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.874179 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k76c5\" (UniqueName: \"kubernetes.io/projected/241b4766-f3d7-4953-845a-6a898cea420e-kube-api-access-k76c5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:28 crc kubenswrapper[4992]: I1004 05:10:28.917812 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:10:29 crc kubenswrapper[4992]: I1004 05:10:29.410805 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69"] Oct 04 05:10:29 crc kubenswrapper[4992]: I1004 05:10:29.461935 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" event={"ID":"241b4766-f3d7-4953-845a-6a898cea420e","Type":"ContainerStarted","Data":"460849f05f4a90075fe8ec757fb8d7fb0774f10dae3b66756b5fb0873675a594"} Oct 04 05:10:30 crc kubenswrapper[4992]: I1004 05:10:30.485044 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" event={"ID":"241b4766-f3d7-4953-845a-6a898cea420e","Type":"ContainerStarted","Data":"aa3fedf4056fc2045d5dddf04bcaa29a12f0653bf80564629bd4c79862d06e26"} Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.640500 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" podStartSLOduration=12.22522971 podStartE2EDuration="12.640481303s" podCreationTimestamp="2025-10-04 05:10:28 +0000 UTC" firstStartedPulling="2025-10-04 05:10:29.405493818 +0000 UTC m=+2263.253169286" lastFinishedPulling="2025-10-04 05:10:29.820745421 +0000 UTC m=+2263.668420879" observedRunningTime="2025-10-04 05:10:30.521613011 +0000 UTC m=+2264.369288499" watchObservedRunningTime="2025-10-04 05:10:40.640481303 +0000 UTC m=+2274.488156771" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.642831 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6w286"] Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.644797 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.656138 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6w286"] Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.780309 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-catalog-content\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.780878 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-utilities\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.780958 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x77vn\" (UniqueName: \"kubernetes.io/projected/d3a24e40-4cfd-446c-b055-c3f6d3867147-kube-api-access-x77vn\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.882486 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-utilities\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.882594 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x77vn\" (UniqueName: \"kubernetes.io/projected/d3a24e40-4cfd-446c-b055-c3f6d3867147-kube-api-access-x77vn\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.882658 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-catalog-content\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.883139 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-utilities\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.883192 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-catalog-content\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.902346 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x77vn\" (UniqueName: \"kubernetes.io/projected/d3a24e40-4cfd-446c-b055-c3f6d3867147-kube-api-access-x77vn\") pod \"redhat-marketplace-6w286\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:40 crc kubenswrapper[4992]: I1004 05:10:40.972901 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:10:41 crc kubenswrapper[4992]: I1004 05:10:41.451712 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6w286"] Oct 04 05:10:41 crc kubenswrapper[4992]: W1004 05:10:41.454927 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3a24e40_4cfd_446c_b055_c3f6d3867147.slice/crio-7762aa61bd38c5676b80e7fb134d7fa722b7d10d22b06fe624791a5d23c58750 WatchSource:0}: Error finding container 7762aa61bd38c5676b80e7fb134d7fa722b7d10d22b06fe624791a5d23c58750: Status 404 returned error can't find the container with id 7762aa61bd38c5676b80e7fb134d7fa722b7d10d22b06fe624791a5d23c58750 Oct 04 05:10:41 crc kubenswrapper[4992]: I1004 05:10:41.575546 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6w286" event={"ID":"d3a24e40-4cfd-446c-b055-c3f6d3867147","Type":"ContainerStarted","Data":"7762aa61bd38c5676b80e7fb134d7fa722b7d10d22b06fe624791a5d23c58750"} Oct 04 05:10:42 crc kubenswrapper[4992]: I1004 05:10:42.585548 4992 generic.go:334] "Generic (PLEG): container finished" podID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerID="4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333" exitCode=0 Oct 04 05:10:42 crc kubenswrapper[4992]: I1004 05:10:42.585628 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6w286" event={"ID":"d3a24e40-4cfd-446c-b055-c3f6d3867147","Type":"ContainerDied","Data":"4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333"} Oct 04 05:10:49 crc kubenswrapper[4992]: I1004 05:10:49.661633 4992 generic.go:334] "Generic (PLEG): container finished" podID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerID="e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98" exitCode=0 Oct 04 05:10:49 crc kubenswrapper[4992]: I1004 05:10:49.661747 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6w286" event={"ID":"d3a24e40-4cfd-446c-b055-c3f6d3867147","Type":"ContainerDied","Data":"e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98"} Oct 04 05:10:52 crc kubenswrapper[4992]: I1004 05:10:52.714133 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6w286" event={"ID":"d3a24e40-4cfd-446c-b055-c3f6d3867147","Type":"ContainerStarted","Data":"dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d"} Oct 04 05:10:52 crc kubenswrapper[4992]: I1004 05:10:52.745226 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6w286" podStartSLOduration=3.347106838 podStartE2EDuration="12.74519447s" podCreationTimestamp="2025-10-04 05:10:40 +0000 UTC" firstStartedPulling="2025-10-04 05:10:42.587087791 +0000 UTC m=+2276.434763269" lastFinishedPulling="2025-10-04 05:10:51.985175403 +0000 UTC m=+2285.832850901" observedRunningTime="2025-10-04 05:10:52.73055074 +0000 UTC m=+2286.578226218" watchObservedRunningTime="2025-10-04 05:10:52.74519447 +0000 UTC m=+2286.592869978" Oct 04 05:11:00 crc kubenswrapper[4992]: I1004 05:11:00.974547 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:11:00 crc kubenswrapper[4992]: I1004 05:11:00.975046 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:11:01 crc kubenswrapper[4992]: I1004 05:11:01.033189 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:11:01 crc kubenswrapper[4992]: I1004 05:11:01.866346 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:11:01 crc kubenswrapper[4992]: I1004 05:11:01.910088 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6w286"] Oct 04 05:11:03 crc kubenswrapper[4992]: I1004 05:11:03.827722 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6w286" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerName="registry-server" containerID="cri-o://dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d" gracePeriod=2 Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.272334 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.383083 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x77vn\" (UniqueName: \"kubernetes.io/projected/d3a24e40-4cfd-446c-b055-c3f6d3867147-kube-api-access-x77vn\") pod \"d3a24e40-4cfd-446c-b055-c3f6d3867147\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.383288 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-catalog-content\") pod \"d3a24e40-4cfd-446c-b055-c3f6d3867147\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.383384 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-utilities\") pod \"d3a24e40-4cfd-446c-b055-c3f6d3867147\" (UID: \"d3a24e40-4cfd-446c-b055-c3f6d3867147\") " Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.384430 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-utilities" (OuterVolumeSpecName: "utilities") pod "d3a24e40-4cfd-446c-b055-c3f6d3867147" (UID: "d3a24e40-4cfd-446c-b055-c3f6d3867147"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.391002 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a24e40-4cfd-446c-b055-c3f6d3867147-kube-api-access-x77vn" (OuterVolumeSpecName: "kube-api-access-x77vn") pod "d3a24e40-4cfd-446c-b055-c3f6d3867147" (UID: "d3a24e40-4cfd-446c-b055-c3f6d3867147"). InnerVolumeSpecName "kube-api-access-x77vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.395736 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3a24e40-4cfd-446c-b055-c3f6d3867147" (UID: "d3a24e40-4cfd-446c-b055-c3f6d3867147"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.485400 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x77vn\" (UniqueName: \"kubernetes.io/projected/d3a24e40-4cfd-446c-b055-c3f6d3867147-kube-api-access-x77vn\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.485442 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.485452 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a24e40-4cfd-446c-b055-c3f6d3867147-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.839987 4992 generic.go:334] "Generic (PLEG): container finished" podID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerID="dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d" exitCode=0 Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.840040 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6w286" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.840079 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6w286" event={"ID":"d3a24e40-4cfd-446c-b055-c3f6d3867147","Type":"ContainerDied","Data":"dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d"} Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.840428 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6w286" event={"ID":"d3a24e40-4cfd-446c-b055-c3f6d3867147","Type":"ContainerDied","Data":"7762aa61bd38c5676b80e7fb134d7fa722b7d10d22b06fe624791a5d23c58750"} Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.840459 4992 scope.go:117] "RemoveContainer" containerID="dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.869440 4992 scope.go:117] "RemoveContainer" containerID="e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.873078 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6w286"] Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.881553 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6w286"] Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.890091 4992 scope.go:117] "RemoveContainer" containerID="4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.940301 4992 scope.go:117] "RemoveContainer" containerID="dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d" Oct 04 05:11:04 crc kubenswrapper[4992]: E1004 05:11:04.940968 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d\": container with ID starting with dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d not found: ID does not exist" containerID="dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.941021 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d"} err="failed to get container status \"dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d\": rpc error: code = NotFound desc = could not find container \"dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d\": container with ID starting with dcbf8fb964da785927527778d7ee8cfbe1a958078aeca25ed75b5354c3344e2d not found: ID does not exist" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.941047 4992 scope.go:117] "RemoveContainer" containerID="e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98" Oct 04 05:11:04 crc kubenswrapper[4992]: E1004 05:11:04.941370 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98\": container with ID starting with e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98 not found: ID does not exist" containerID="e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.941392 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98"} err="failed to get container status \"e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98\": rpc error: code = NotFound desc = could not find container \"e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98\": container with ID starting with e5a9078a5ea1bf57a793e9940cf17d801b2d459542424f03bb68478a0ec83a98 not found: ID does not exist" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.941406 4992 scope.go:117] "RemoveContainer" containerID="4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333" Oct 04 05:11:04 crc kubenswrapper[4992]: E1004 05:11:04.941726 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333\": container with ID starting with 4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333 not found: ID does not exist" containerID="4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333" Oct 04 05:11:04 crc kubenswrapper[4992]: I1004 05:11:04.941751 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333"} err="failed to get container status \"4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333\": rpc error: code = NotFound desc = could not find container \"4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333\": container with ID starting with 4538e2f7d6849f0d27fdda118e48146db9c488f51f01aedf020b467807485333 not found: ID does not exist" Oct 04 05:11:06 crc kubenswrapper[4992]: I1004 05:11:06.330232 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" path="/var/lib/kubelet/pods/d3a24e40-4cfd-446c-b055-c3f6d3867147/volumes" Oct 04 05:11:16 crc kubenswrapper[4992]: I1004 05:11:16.937749 4992 generic.go:334] "Generic (PLEG): container finished" podID="241b4766-f3d7-4953-845a-6a898cea420e" containerID="aa3fedf4056fc2045d5dddf04bcaa29a12f0653bf80564629bd4c79862d06e26" exitCode=0 Oct 04 05:11:16 crc kubenswrapper[4992]: I1004 05:11:16.937868 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" event={"ID":"241b4766-f3d7-4953-845a-6a898cea420e","Type":"ContainerDied","Data":"aa3fedf4056fc2045d5dddf04bcaa29a12f0653bf80564629bd4c79862d06e26"} Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.351444 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.449131 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"241b4766-f3d7-4953-845a-6a898cea420e\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.449227 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k76c5\" (UniqueName: \"kubernetes.io/projected/241b4766-f3d7-4953-845a-6a898cea420e-kube-api-access-k76c5\") pod \"241b4766-f3d7-4953-845a-6a898cea420e\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.449576 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-metadata-combined-ca-bundle\") pod \"241b4766-f3d7-4953-845a-6a898cea420e\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.449618 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-nova-metadata-neutron-config-0\") pod \"241b4766-f3d7-4953-845a-6a898cea420e\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.449648 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-ssh-key\") pod \"241b4766-f3d7-4953-845a-6a898cea420e\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.449746 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-inventory\") pod \"241b4766-f3d7-4953-845a-6a898cea420e\" (UID: \"241b4766-f3d7-4953-845a-6a898cea420e\") " Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.455163 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/241b4766-f3d7-4953-845a-6a898cea420e-kube-api-access-k76c5" (OuterVolumeSpecName: "kube-api-access-k76c5") pod "241b4766-f3d7-4953-845a-6a898cea420e" (UID: "241b4766-f3d7-4953-845a-6a898cea420e"). InnerVolumeSpecName "kube-api-access-k76c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.456307 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "241b4766-f3d7-4953-845a-6a898cea420e" (UID: "241b4766-f3d7-4953-845a-6a898cea420e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.477337 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "241b4766-f3d7-4953-845a-6a898cea420e" (UID: "241b4766-f3d7-4953-845a-6a898cea420e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.480855 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-inventory" (OuterVolumeSpecName: "inventory") pod "241b4766-f3d7-4953-845a-6a898cea420e" (UID: "241b4766-f3d7-4953-845a-6a898cea420e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.482935 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "241b4766-f3d7-4953-845a-6a898cea420e" (UID: "241b4766-f3d7-4953-845a-6a898cea420e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.483900 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "241b4766-f3d7-4953-845a-6a898cea420e" (UID: "241b4766-f3d7-4953-845a-6a898cea420e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.555031 4992 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.555098 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.555117 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.555131 4992 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.555176 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k76c5\" (UniqueName: \"kubernetes.io/projected/241b4766-f3d7-4953-845a-6a898cea420e-kube-api-access-k76c5\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.555193 4992 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241b4766-f3d7-4953-845a-6a898cea420e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.960519 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" event={"ID":"241b4766-f3d7-4953-845a-6a898cea420e","Type":"ContainerDied","Data":"460849f05f4a90075fe8ec757fb8d7fb0774f10dae3b66756b5fb0873675a594"} Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.961043 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="460849f05f4a90075fe8ec757fb8d7fb0774f10dae3b66756b5fb0873675a594" Oct 04 05:11:18 crc kubenswrapper[4992]: I1004 05:11:18.960957 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.036602 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm"] Oct 04 05:11:19 crc kubenswrapper[4992]: E1004 05:11:19.037031 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerName="registry-server" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.037052 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerName="registry-server" Oct 04 05:11:19 crc kubenswrapper[4992]: E1004 05:11:19.037072 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerName="extract-utilities" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.037080 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerName="extract-utilities" Oct 04 05:11:19 crc kubenswrapper[4992]: E1004 05:11:19.037098 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241b4766-f3d7-4953-845a-6a898cea420e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.037108 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="241b4766-f3d7-4953-845a-6a898cea420e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 05:11:19 crc kubenswrapper[4992]: E1004 05:11:19.037120 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerName="extract-content" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.037126 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerName="extract-content" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.037311 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="241b4766-f3d7-4953-845a-6a898cea420e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.037336 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a24e40-4cfd-446c-b055-c3f6d3867147" containerName="registry-server" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.038004 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.042261 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.042478 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.042588 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.042810 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.042930 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.051346 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm"] Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.063714 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.063824 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.063851 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn2jb\" (UniqueName: \"kubernetes.io/projected/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-kube-api-access-wn2jb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.063883 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.063912 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.164974 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.165307 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.165509 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.165608 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn2jb\" (UniqueName: \"kubernetes.io/projected/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-kube-api-access-wn2jb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.165731 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.169767 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.170578 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.170708 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.171709 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.186387 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn2jb\" (UniqueName: \"kubernetes.io/projected/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-kube-api-access-wn2jb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.358819 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:11:19 crc kubenswrapper[4992]: I1004 05:11:19.982972 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm"] Oct 04 05:11:19 crc kubenswrapper[4992]: W1004 05:11:19.990766 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41be67bd_77e6_4b0b_bf54_9c3ffb228ea9.slice/crio-603dee785dbbd1b5fe62bdc08dd8f99183a5c023ddf8e85d8e951f1f69239aa3 WatchSource:0}: Error finding container 603dee785dbbd1b5fe62bdc08dd8f99183a5c023ddf8e85d8e951f1f69239aa3: Status 404 returned error can't find the container with id 603dee785dbbd1b5fe62bdc08dd8f99183a5c023ddf8e85d8e951f1f69239aa3 Oct 04 05:11:20 crc kubenswrapper[4992]: I1004 05:11:20.978025 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" event={"ID":"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9","Type":"ContainerStarted","Data":"685dc6b284a84fdc62432a8581e32299fb3cdaa6cc3232830d1b91e7e0371a18"} Oct 04 05:11:20 crc kubenswrapper[4992]: I1004 05:11:20.978945 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" event={"ID":"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9","Type":"ContainerStarted","Data":"603dee785dbbd1b5fe62bdc08dd8f99183a5c023ddf8e85d8e951f1f69239aa3"} Oct 04 05:11:21 crc kubenswrapper[4992]: I1004 05:11:21.000241 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" podStartSLOduration=1.284139554 podStartE2EDuration="2.000217689s" podCreationTimestamp="2025-10-04 05:11:19 +0000 UTC" firstStartedPulling="2025-10-04 05:11:19.993974676 +0000 UTC m=+2313.841650144" lastFinishedPulling="2025-10-04 05:11:20.710052811 +0000 UTC m=+2314.557728279" observedRunningTime="2025-10-04 05:11:20.994110536 +0000 UTC m=+2314.841786064" watchObservedRunningTime="2025-10-04 05:11:21.000217689 +0000 UTC m=+2314.847893177" Oct 04 05:12:20 crc kubenswrapper[4992]: I1004 05:12:20.894202 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:12:20 crc kubenswrapper[4992]: I1004 05:12:20.894779 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:12:50 crc kubenswrapper[4992]: I1004 05:12:50.894591 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:12:50 crc kubenswrapper[4992]: I1004 05:12:50.895202 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:13:20 crc kubenswrapper[4992]: I1004 05:13:20.894704 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:13:20 crc kubenswrapper[4992]: I1004 05:13:20.895275 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:13:20 crc kubenswrapper[4992]: I1004 05:13:20.895330 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 05:13:20 crc kubenswrapper[4992]: I1004 05:13:20.896138 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:13:20 crc kubenswrapper[4992]: I1004 05:13:20.896205 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" gracePeriod=600 Oct 04 05:13:21 crc kubenswrapper[4992]: I1004 05:13:21.127786 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" exitCode=0 Oct 04 05:13:21 crc kubenswrapper[4992]: I1004 05:13:21.127829 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505"} Oct 04 05:13:21 crc kubenswrapper[4992]: I1004 05:13:21.127876 4992 scope.go:117] "RemoveContainer" containerID="3bca95cda3843b59172df471f72becf1de18ceb48613d521c6b0ccd2ca05ae02" Oct 04 05:13:21 crc kubenswrapper[4992]: E1004 05:13:21.588053 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:13:22 crc kubenswrapper[4992]: I1004 05:13:22.146032 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:13:22 crc kubenswrapper[4992]: E1004 05:13:22.146503 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:13:36 crc kubenswrapper[4992]: I1004 05:13:36.319673 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:13:36 crc kubenswrapper[4992]: E1004 05:13:36.320663 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:13:47 crc kubenswrapper[4992]: I1004 05:13:47.319756 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:13:47 crc kubenswrapper[4992]: E1004 05:13:47.320498 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:14:02 crc kubenswrapper[4992]: I1004 05:14:02.319795 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:14:02 crc kubenswrapper[4992]: E1004 05:14:02.320467 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:14:17 crc kubenswrapper[4992]: I1004 05:14:17.319792 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:14:17 crc kubenswrapper[4992]: E1004 05:14:17.320674 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:14:32 crc kubenswrapper[4992]: I1004 05:14:32.319896 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:14:32 crc kubenswrapper[4992]: E1004 05:14:32.321024 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:14:45 crc kubenswrapper[4992]: I1004 05:14:45.319308 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:14:45 crc kubenswrapper[4992]: E1004 05:14:45.320079 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:14:58 crc kubenswrapper[4992]: I1004 05:14:58.319576 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:14:58 crc kubenswrapper[4992]: E1004 05:14:58.320452 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.162657 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6"] Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.164837 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.166894 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.167655 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.174217 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6"] Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.209922 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a54979a-8679-4a23-a955-828eafa964e3-config-volume\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.210335 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlsgz\" (UniqueName: \"kubernetes.io/projected/4a54979a-8679-4a23-a955-828eafa964e3-kube-api-access-mlsgz\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.210671 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a54979a-8679-4a23-a955-828eafa964e3-secret-volume\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.313748 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlsgz\" (UniqueName: \"kubernetes.io/projected/4a54979a-8679-4a23-a955-828eafa964e3-kube-api-access-mlsgz\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.313925 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a54979a-8679-4a23-a955-828eafa964e3-secret-volume\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.314001 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a54979a-8679-4a23-a955-828eafa964e3-config-volume\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.315412 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a54979a-8679-4a23-a955-828eafa964e3-config-volume\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.325613 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a54979a-8679-4a23-a955-828eafa964e3-secret-volume\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.343542 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlsgz\" (UniqueName: \"kubernetes.io/projected/4a54979a-8679-4a23-a955-828eafa964e3-kube-api-access-mlsgz\") pod \"collect-profiles-29325915-7qcd6\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.485712 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:00 crc kubenswrapper[4992]: I1004 05:15:00.953799 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6"] Oct 04 05:15:01 crc kubenswrapper[4992]: I1004 05:15:01.074764 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" event={"ID":"4a54979a-8679-4a23-a955-828eafa964e3","Type":"ContainerStarted","Data":"1324ae71dee8b86653179a1f005e1d88caf589d80b4d5123794d03c461ed832c"} Oct 04 05:15:02 crc kubenswrapper[4992]: I1004 05:15:02.085276 4992 generic.go:334] "Generic (PLEG): container finished" podID="4a54979a-8679-4a23-a955-828eafa964e3" containerID="e54850b31718b959786db0ab44cf99143fe9a27027824c88c17ad630e8bf4d0a" exitCode=0 Oct 04 05:15:02 crc kubenswrapper[4992]: I1004 05:15:02.085382 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" event={"ID":"4a54979a-8679-4a23-a955-828eafa964e3","Type":"ContainerDied","Data":"e54850b31718b959786db0ab44cf99143fe9a27027824c88c17ad630e8bf4d0a"} Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.449492 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.473684 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlsgz\" (UniqueName: \"kubernetes.io/projected/4a54979a-8679-4a23-a955-828eafa964e3-kube-api-access-mlsgz\") pod \"4a54979a-8679-4a23-a955-828eafa964e3\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.473731 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a54979a-8679-4a23-a955-828eafa964e3-secret-volume\") pod \"4a54979a-8679-4a23-a955-828eafa964e3\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.484122 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a54979a-8679-4a23-a955-828eafa964e3-kube-api-access-mlsgz" (OuterVolumeSpecName: "kube-api-access-mlsgz") pod "4a54979a-8679-4a23-a955-828eafa964e3" (UID: "4a54979a-8679-4a23-a955-828eafa964e3"). InnerVolumeSpecName "kube-api-access-mlsgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.486049 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a54979a-8679-4a23-a955-828eafa964e3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4a54979a-8679-4a23-a955-828eafa964e3" (UID: "4a54979a-8679-4a23-a955-828eafa964e3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.575494 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a54979a-8679-4a23-a955-828eafa964e3-config-volume\") pod \"4a54979a-8679-4a23-a955-828eafa964e3\" (UID: \"4a54979a-8679-4a23-a955-828eafa964e3\") " Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.576056 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlsgz\" (UniqueName: \"kubernetes.io/projected/4a54979a-8679-4a23-a955-828eafa964e3-kube-api-access-mlsgz\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.576129 4992 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a54979a-8679-4a23-a955-828eafa964e3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.576318 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a54979a-8679-4a23-a955-828eafa964e3-config-volume" (OuterVolumeSpecName: "config-volume") pod "4a54979a-8679-4a23-a955-828eafa964e3" (UID: "4a54979a-8679-4a23-a955-828eafa964e3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:15:03 crc kubenswrapper[4992]: I1004 05:15:03.677563 4992 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a54979a-8679-4a23-a955-828eafa964e3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:04 crc kubenswrapper[4992]: I1004 05:15:04.102666 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" event={"ID":"4a54979a-8679-4a23-a955-828eafa964e3","Type":"ContainerDied","Data":"1324ae71dee8b86653179a1f005e1d88caf589d80b4d5123794d03c461ed832c"} Oct 04 05:15:04 crc kubenswrapper[4992]: I1004 05:15:04.102714 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1324ae71dee8b86653179a1f005e1d88caf589d80b4d5123794d03c461ed832c" Oct 04 05:15:04 crc kubenswrapper[4992]: I1004 05:15:04.103045 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325915-7qcd6" Oct 04 05:15:04 crc kubenswrapper[4992]: I1004 05:15:04.564254 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n"] Oct 04 05:15:04 crc kubenswrapper[4992]: I1004 05:15:04.573928 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325870-cvw6n"] Oct 04 05:15:06 crc kubenswrapper[4992]: I1004 05:15:06.330117 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24abe58b-4b74-4207-ba8f-1b16c22617cc" path="/var/lib/kubelet/pods/24abe58b-4b74-4207-ba8f-1b16c22617cc/volumes" Oct 04 05:15:10 crc kubenswrapper[4992]: I1004 05:15:10.320060 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:15:10 crc kubenswrapper[4992]: E1004 05:15:10.321685 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:15:12 crc kubenswrapper[4992]: I1004 05:15:12.201855 4992 scope.go:117] "RemoveContainer" containerID="dfc12adf892ce475e076b6342b9a6c6683f9436b40f71461d573eb93665c120b" Oct 04 05:15:21 crc kubenswrapper[4992]: I1004 05:15:21.319145 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:15:21 crc kubenswrapper[4992]: E1004 05:15:21.321072 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:15:29 crc kubenswrapper[4992]: I1004 05:15:29.358129 4992 generic.go:334] "Generic (PLEG): container finished" podID="41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" containerID="685dc6b284a84fdc62432a8581e32299fb3cdaa6cc3232830d1b91e7e0371a18" exitCode=0 Oct 04 05:15:29 crc kubenswrapper[4992]: I1004 05:15:29.358213 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" event={"ID":"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9","Type":"ContainerDied","Data":"685dc6b284a84fdc62432a8581e32299fb3cdaa6cc3232830d1b91e7e0371a18"} Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.817436 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.826151 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn2jb\" (UniqueName: \"kubernetes.io/projected/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-kube-api-access-wn2jb\") pod \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.826312 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-combined-ca-bundle\") pod \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.826449 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-inventory\") pod \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.826481 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-secret-0\") pod \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.826529 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-ssh-key\") pod \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\" (UID: \"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9\") " Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.836183 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" (UID: "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.836240 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-kube-api-access-wn2jb" (OuterVolumeSpecName: "kube-api-access-wn2jb") pod "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" (UID: "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9"). InnerVolumeSpecName "kube-api-access-wn2jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.861853 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-inventory" (OuterVolumeSpecName: "inventory") pod "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" (UID: "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.869647 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" (UID: "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.870292 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" (UID: "41be67bd-77e6-4b0b-bf54-9c3ffb228ea9"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.929654 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn2jb\" (UniqueName: \"kubernetes.io/projected/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-kube-api-access-wn2jb\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.929693 4992 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.929703 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.929715 4992 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:30 crc kubenswrapper[4992]: I1004 05:15:30.929724 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41be67bd-77e6-4b0b-bf54-9c3ffb228ea9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.377024 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" event={"ID":"41be67bd-77e6-4b0b-bf54-9c3ffb228ea9","Type":"ContainerDied","Data":"603dee785dbbd1b5fe62bdc08dd8f99183a5c023ddf8e85d8e951f1f69239aa3"} Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.377095 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="603dee785dbbd1b5fe62bdc08dd8f99183a5c023ddf8e85d8e951f1f69239aa3" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.377101 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.457226 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2"] Oct 04 05:15:31 crc kubenswrapper[4992]: E1004 05:15:31.458071 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.458092 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 05:15:31 crc kubenswrapper[4992]: E1004 05:15:31.458105 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a54979a-8679-4a23-a955-828eafa964e3" containerName="collect-profiles" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.458114 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a54979a-8679-4a23-a955-828eafa964e3" containerName="collect-profiles" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.458318 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="41be67bd-77e6-4b0b-bf54-9c3ffb228ea9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.458354 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a54979a-8679-4a23-a955-828eafa964e3" containerName="collect-profiles" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.459111 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.472009 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.472015 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.472112 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.472435 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.472455 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.472524 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.472831 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.473487 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2"] Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.541866 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.541911 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.542122 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.542259 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a29da406-d169-4c98-ac72-45380f713377-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.542347 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bmbc\" (UniqueName: \"kubernetes.io/projected/a29da406-d169-4c98-ac72-45380f713377-kube-api-access-4bmbc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.542535 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.542679 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.542815 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.542880 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.643880 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a29da406-d169-4c98-ac72-45380f713377-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.643940 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bmbc\" (UniqueName: \"kubernetes.io/projected/a29da406-d169-4c98-ac72-45380f713377-kube-api-access-4bmbc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.644002 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.644046 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.644083 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.644122 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.644177 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.644207 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.644248 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.645170 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a29da406-d169-4c98-ac72-45380f713377-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.648827 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.649244 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.649881 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.650228 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.654998 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.655576 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.655966 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.663908 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bmbc\" (UniqueName: \"kubernetes.io/projected/a29da406-d169-4c98-ac72-45380f713377-kube-api-access-4bmbc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8dcg2\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:31 crc kubenswrapper[4992]: I1004 05:15:31.782315 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:15:32 crc kubenswrapper[4992]: I1004 05:15:32.296185 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2"] Oct 04 05:15:32 crc kubenswrapper[4992]: I1004 05:15:32.304574 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:15:32 crc kubenswrapper[4992]: I1004 05:15:32.385550 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" event={"ID":"a29da406-d169-4c98-ac72-45380f713377","Type":"ContainerStarted","Data":"3ba9ee05fbf2574b3531a2a0284c877cc737384fe424cb4f8186a8549afcd6ed"} Oct 04 05:15:33 crc kubenswrapper[4992]: I1004 05:15:33.414903 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" event={"ID":"a29da406-d169-4c98-ac72-45380f713377","Type":"ContainerStarted","Data":"e74c2c753c2313017b6ee2da1641f9fdbe4f6975dbc4f9b09c998f748d1e0eff"} Oct 04 05:15:33 crc kubenswrapper[4992]: I1004 05:15:33.436006 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" podStartSLOduration=2.006054957 podStartE2EDuration="2.435984013s" podCreationTimestamp="2025-10-04 05:15:31 +0000 UTC" firstStartedPulling="2025-10-04 05:15:32.304317354 +0000 UTC m=+2566.151992822" lastFinishedPulling="2025-10-04 05:15:32.73424641 +0000 UTC m=+2566.581921878" observedRunningTime="2025-10-04 05:15:33.430014543 +0000 UTC m=+2567.277690031" watchObservedRunningTime="2025-10-04 05:15:33.435984013 +0000 UTC m=+2567.283659481" Oct 04 05:15:35 crc kubenswrapper[4992]: I1004 05:15:35.319099 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:15:35 crc kubenswrapper[4992]: E1004 05:15:35.319668 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:15:48 crc kubenswrapper[4992]: I1004 05:15:48.321044 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:15:48 crc kubenswrapper[4992]: E1004 05:15:48.321853 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:16:03 crc kubenswrapper[4992]: I1004 05:16:03.319742 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:16:03 crc kubenswrapper[4992]: E1004 05:16:03.320647 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:16:12 crc kubenswrapper[4992]: I1004 05:16:12.282441 4992 scope.go:117] "RemoveContainer" containerID="a97006b34096e177e3b77ea78eaa3e729ca9beaf3a841f09d13b3b1553188921" Oct 04 05:16:12 crc kubenswrapper[4992]: I1004 05:16:12.317465 4992 scope.go:117] "RemoveContainer" containerID="6f6cd2fcb6d9602d858a5e2725f331e4e385a0a8a8cdc8417ae51514412b25c5" Oct 04 05:16:12 crc kubenswrapper[4992]: I1004 05:16:12.376269 4992 scope.go:117] "RemoveContainer" containerID="8b1569a1a9d71915ee550bece19c6c174925e72a1f695e955d9be0c51656b727" Oct 04 05:16:14 crc kubenswrapper[4992]: I1004 05:16:14.334513 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:16:14 crc kubenswrapper[4992]: E1004 05:16:14.334947 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:16:28 crc kubenswrapper[4992]: I1004 05:16:28.320056 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:16:28 crc kubenswrapper[4992]: E1004 05:16:28.322491 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:16:41 crc kubenswrapper[4992]: I1004 05:16:41.319487 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:16:41 crc kubenswrapper[4992]: E1004 05:16:41.320380 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:16:54 crc kubenswrapper[4992]: I1004 05:16:54.332974 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:16:54 crc kubenswrapper[4992]: E1004 05:16:54.333813 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:17:05 crc kubenswrapper[4992]: I1004 05:17:05.319867 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:17:05 crc kubenswrapper[4992]: E1004 05:17:05.320791 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:17:18 crc kubenswrapper[4992]: I1004 05:17:18.319311 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:17:18 crc kubenswrapper[4992]: E1004 05:17:18.319982 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:17:32 crc kubenswrapper[4992]: I1004 05:17:32.319668 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:17:32 crc kubenswrapper[4992]: E1004 05:17:32.320422 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:17:44 crc kubenswrapper[4992]: I1004 05:17:44.324696 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:17:44 crc kubenswrapper[4992]: E1004 05:17:44.325435 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:17:56 crc kubenswrapper[4992]: I1004 05:17:56.319805 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:17:56 crc kubenswrapper[4992]: E1004 05:17:56.320765 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:18:08 crc kubenswrapper[4992]: I1004 05:18:08.319016 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:18:08 crc kubenswrapper[4992]: E1004 05:18:08.319751 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:18:19 crc kubenswrapper[4992]: I1004 05:18:19.319966 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:18:19 crc kubenswrapper[4992]: E1004 05:18:19.320740 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:18:34 crc kubenswrapper[4992]: I1004 05:18:34.326721 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:18:35 crc kubenswrapper[4992]: I1004 05:18:35.069197 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"46df65f6a9ec684c862b0e61ea079465d75bc9be637895142632bb33563718a4"} Oct 04 05:18:40 crc kubenswrapper[4992]: I1004 05:18:40.111335 4992 generic.go:334] "Generic (PLEG): container finished" podID="a29da406-d169-4c98-ac72-45380f713377" containerID="e74c2c753c2313017b6ee2da1641f9fdbe4f6975dbc4f9b09c998f748d1e0eff" exitCode=0 Oct 04 05:18:40 crc kubenswrapper[4992]: I1004 05:18:40.111390 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" event={"ID":"a29da406-d169-4c98-ac72-45380f713377","Type":"ContainerDied","Data":"e74c2c753c2313017b6ee2da1641f9fdbe4f6975dbc4f9b09c998f748d1e0eff"} Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.537287 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.573544 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-0\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.573603 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-combined-ca-bundle\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.573701 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a29da406-d169-4c98-ac72-45380f713377-nova-extra-config-0\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.573753 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-1\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.573785 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-ssh-key\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.573849 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-1\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.573921 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-0\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.573965 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bmbc\" (UniqueName: \"kubernetes.io/projected/a29da406-d169-4c98-ac72-45380f713377-kube-api-access-4bmbc\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.574033 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-inventory\") pod \"a29da406-d169-4c98-ac72-45380f713377\" (UID: \"a29da406-d169-4c98-ac72-45380f713377\") " Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.580929 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.582323 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29da406-d169-4c98-ac72-45380f713377-kube-api-access-4bmbc" (OuterVolumeSpecName: "kube-api-access-4bmbc") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "kube-api-access-4bmbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.610027 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.613714 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.614201 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a29da406-d169-4c98-ac72-45380f713377-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.614521 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-inventory" (OuterVolumeSpecName: "inventory") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.616480 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.616615 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.626619 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "a29da406-d169-4c98-ac72-45380f713377" (UID: "a29da406-d169-4c98-ac72-45380f713377"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.678963 4992 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.679226 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bmbc\" (UniqueName: \"kubernetes.io/projected/a29da406-d169-4c98-ac72-45380f713377-kube-api-access-4bmbc\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.679285 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.679304 4992 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.679316 4992 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.679449 4992 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a29da406-d169-4c98-ac72-45380f713377-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.679480 4992 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.679493 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:41 crc kubenswrapper[4992]: I1004 05:18:41.679526 4992 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a29da406-d169-4c98-ac72-45380f713377-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.129866 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" event={"ID":"a29da406-d169-4c98-ac72-45380f713377","Type":"ContainerDied","Data":"3ba9ee05fbf2574b3531a2a0284c877cc737384fe424cb4f8186a8549afcd6ed"} Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.129910 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ba9ee05fbf2574b3531a2a0284c877cc737384fe424cb4f8186a8549afcd6ed" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.129922 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8dcg2" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.225725 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs"] Oct 04 05:18:42 crc kubenswrapper[4992]: E1004 05:18:42.226197 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29da406-d169-4c98-ac72-45380f713377" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.226220 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29da406-d169-4c98-ac72-45380f713377" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.226466 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="a29da406-d169-4c98-ac72-45380f713377" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.227229 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.230852 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.231213 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.231480 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.232205 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.232406 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-d5mzk" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.288232 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.288318 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.288399 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.288427 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-528nz\" (UniqueName: \"kubernetes.io/projected/8e39e2be-f116-452e-ac17-3b7337d80f5d-kube-api-access-528nz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.288488 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.288532 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.288682 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.290241 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs"] Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.389834 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.389904 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-528nz\" (UniqueName: \"kubernetes.io/projected/8e39e2be-f116-452e-ac17-3b7337d80f5d-kube-api-access-528nz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.389956 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.389991 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.392415 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.392566 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.393274 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.402255 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.403815 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.404487 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.404904 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.405297 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.406674 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.428265 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-528nz\" (UniqueName: \"kubernetes.io/projected/8e39e2be-f116-452e-ac17-3b7337d80f5d-kube-api-access-528nz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:42 crc kubenswrapper[4992]: I1004 05:18:42.543074 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:18:43 crc kubenswrapper[4992]: I1004 05:18:43.065215 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs"] Oct 04 05:18:43 crc kubenswrapper[4992]: I1004 05:18:43.144102 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" event={"ID":"8e39e2be-f116-452e-ac17-3b7337d80f5d","Type":"ContainerStarted","Data":"55ef96937f88b003c31e3faa6153bccc97ac2ac490c0190329481617e8cfa662"} Oct 04 05:18:44 crc kubenswrapper[4992]: I1004 05:18:44.155398 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" event={"ID":"8e39e2be-f116-452e-ac17-3b7337d80f5d","Type":"ContainerStarted","Data":"00cb65443ce41cfd34dc22d6ec5b3e088306e686a275f56c1ce1a835871e5fc2"} Oct 04 05:18:44 crc kubenswrapper[4992]: I1004 05:18:44.177935 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" podStartSLOduration=1.741296937 podStartE2EDuration="2.177912509s" podCreationTimestamp="2025-10-04 05:18:42 +0000 UTC" firstStartedPulling="2025-10-04 05:18:43.074047736 +0000 UTC m=+2756.921723204" lastFinishedPulling="2025-10-04 05:18:43.510663308 +0000 UTC m=+2757.358338776" observedRunningTime="2025-10-04 05:18:44.169957227 +0000 UTC m=+2758.017632695" watchObservedRunningTime="2025-10-04 05:18:44.177912509 +0000 UTC m=+2758.025588007" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.191697 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tmb5s"] Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.195318 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.222644 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tmb5s"] Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.283735 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpsg5\" (UniqueName: \"kubernetes.io/projected/f391f4ad-e25f-49a1-b337-3e6fc34b358d-kube-api-access-fpsg5\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.283800 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-catalog-content\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.283847 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-utilities\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.386235 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpsg5\" (UniqueName: \"kubernetes.io/projected/f391f4ad-e25f-49a1-b337-3e6fc34b358d-kube-api-access-fpsg5\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.386290 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-catalog-content\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.386333 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-utilities\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.387697 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-utilities\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.387947 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-catalog-content\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.411096 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpsg5\" (UniqueName: \"kubernetes.io/projected/f391f4ad-e25f-49a1-b337-3e6fc34b358d-kube-api-access-fpsg5\") pod \"redhat-operators-tmb5s\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:37 crc kubenswrapper[4992]: I1004 05:19:37.536873 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:38 crc kubenswrapper[4992]: W1004 05:19:38.013467 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf391f4ad_e25f_49a1_b337_3e6fc34b358d.slice/crio-ade4946362dc1e84773bb4e100cd70f985aabf756f235799fd25bb23f5c7ea4b WatchSource:0}: Error finding container ade4946362dc1e84773bb4e100cd70f985aabf756f235799fd25bb23f5c7ea4b: Status 404 returned error can't find the container with id ade4946362dc1e84773bb4e100cd70f985aabf756f235799fd25bb23f5c7ea4b Oct 04 05:19:38 crc kubenswrapper[4992]: I1004 05:19:38.017296 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tmb5s"] Oct 04 05:19:38 crc kubenswrapper[4992]: I1004 05:19:38.672700 4992 generic.go:334] "Generic (PLEG): container finished" podID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerID="916675902fc3aba5c181379f64f22d0a2a12e6528913d6eab1219844af7033f5" exitCode=0 Oct 04 05:19:38 crc kubenswrapper[4992]: I1004 05:19:38.672781 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmb5s" event={"ID":"f391f4ad-e25f-49a1-b337-3e6fc34b358d","Type":"ContainerDied","Data":"916675902fc3aba5c181379f64f22d0a2a12e6528913d6eab1219844af7033f5"} Oct 04 05:19:38 crc kubenswrapper[4992]: I1004 05:19:38.673055 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmb5s" event={"ID":"f391f4ad-e25f-49a1-b337-3e6fc34b358d","Type":"ContainerStarted","Data":"ade4946362dc1e84773bb4e100cd70f985aabf756f235799fd25bb23f5c7ea4b"} Oct 04 05:19:40 crc kubenswrapper[4992]: I1004 05:19:40.691522 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmb5s" event={"ID":"f391f4ad-e25f-49a1-b337-3e6fc34b358d","Type":"ContainerStarted","Data":"f05690913c415ef85903deacba101d1d5cc79598aa0842eb4bc0e89f32cbbf32"} Oct 04 05:19:41 crc kubenswrapper[4992]: I1004 05:19:41.703846 4992 generic.go:334] "Generic (PLEG): container finished" podID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerID="f05690913c415ef85903deacba101d1d5cc79598aa0842eb4bc0e89f32cbbf32" exitCode=0 Oct 04 05:19:41 crc kubenswrapper[4992]: I1004 05:19:41.703966 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmb5s" event={"ID":"f391f4ad-e25f-49a1-b337-3e6fc34b358d","Type":"ContainerDied","Data":"f05690913c415ef85903deacba101d1d5cc79598aa0842eb4bc0e89f32cbbf32"} Oct 04 05:19:44 crc kubenswrapper[4992]: I1004 05:19:44.750767 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmb5s" event={"ID":"f391f4ad-e25f-49a1-b337-3e6fc34b358d","Type":"ContainerStarted","Data":"570e5568f1c10c25a3d9566187a706407c99b964acd7633b7953ef359bd7dcb3"} Oct 04 05:19:44 crc kubenswrapper[4992]: I1004 05:19:44.773613 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tmb5s" podStartSLOduration=2.862306941 podStartE2EDuration="7.773594128s" podCreationTimestamp="2025-10-04 05:19:37 +0000 UTC" firstStartedPulling="2025-10-04 05:19:38.674802834 +0000 UTC m=+2812.522478322" lastFinishedPulling="2025-10-04 05:19:43.586090041 +0000 UTC m=+2817.433765509" observedRunningTime="2025-10-04 05:19:44.767501216 +0000 UTC m=+2818.615176684" watchObservedRunningTime="2025-10-04 05:19:44.773594128 +0000 UTC m=+2818.621269596" Oct 04 05:19:47 crc kubenswrapper[4992]: I1004 05:19:47.537980 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:47 crc kubenswrapper[4992]: I1004 05:19:47.538339 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:48 crc kubenswrapper[4992]: I1004 05:19:48.588128 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tmb5s" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="registry-server" probeResult="failure" output=< Oct 04 05:19:48 crc kubenswrapper[4992]: timeout: failed to connect service ":50051" within 1s Oct 04 05:19:48 crc kubenswrapper[4992]: > Oct 04 05:19:57 crc kubenswrapper[4992]: I1004 05:19:57.600770 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:57 crc kubenswrapper[4992]: I1004 05:19:57.657967 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:19:57 crc kubenswrapper[4992]: I1004 05:19:57.851443 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tmb5s"] Oct 04 05:19:58 crc kubenswrapper[4992]: I1004 05:19:58.893803 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tmb5s" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="registry-server" containerID="cri-o://570e5568f1c10c25a3d9566187a706407c99b964acd7633b7953ef359bd7dcb3" gracePeriod=2 Oct 04 05:19:59 crc kubenswrapper[4992]: I1004 05:19:59.908658 4992 generic.go:334] "Generic (PLEG): container finished" podID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerID="570e5568f1c10c25a3d9566187a706407c99b964acd7633b7953ef359bd7dcb3" exitCode=0 Oct 04 05:19:59 crc kubenswrapper[4992]: I1004 05:19:59.908753 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmb5s" event={"ID":"f391f4ad-e25f-49a1-b337-3e6fc34b358d","Type":"ContainerDied","Data":"570e5568f1c10c25a3d9566187a706407c99b964acd7633b7953ef359bd7dcb3"} Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.087711 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.241111 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-catalog-content\") pod \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.241222 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpsg5\" (UniqueName: \"kubernetes.io/projected/f391f4ad-e25f-49a1-b337-3e6fc34b358d-kube-api-access-fpsg5\") pod \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.242911 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-utilities\") pod \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\" (UID: \"f391f4ad-e25f-49a1-b337-3e6fc34b358d\") " Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.244549 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-utilities" (OuterVolumeSpecName: "utilities") pod "f391f4ad-e25f-49a1-b337-3e6fc34b358d" (UID: "f391f4ad-e25f-49a1-b337-3e6fc34b358d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.245074 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.248045 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f391f4ad-e25f-49a1-b337-3e6fc34b358d-kube-api-access-fpsg5" (OuterVolumeSpecName: "kube-api-access-fpsg5") pod "f391f4ad-e25f-49a1-b337-3e6fc34b358d" (UID: "f391f4ad-e25f-49a1-b337-3e6fc34b358d"). InnerVolumeSpecName "kube-api-access-fpsg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.347325 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpsg5\" (UniqueName: \"kubernetes.io/projected/f391f4ad-e25f-49a1-b337-3e6fc34b358d-kube-api-access-fpsg5\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.355145 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f391f4ad-e25f-49a1-b337-3e6fc34b358d" (UID: "f391f4ad-e25f-49a1-b337-3e6fc34b358d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.449761 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f391f4ad-e25f-49a1-b337-3e6fc34b358d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.920584 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tmb5s" event={"ID":"f391f4ad-e25f-49a1-b337-3e6fc34b358d","Type":"ContainerDied","Data":"ade4946362dc1e84773bb4e100cd70f985aabf756f235799fd25bb23f5c7ea4b"} Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.920627 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tmb5s" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.920644 4992 scope.go:117] "RemoveContainer" containerID="570e5568f1c10c25a3d9566187a706407c99b964acd7633b7953ef359bd7dcb3" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.952561 4992 scope.go:117] "RemoveContainer" containerID="f05690913c415ef85903deacba101d1d5cc79598aa0842eb4bc0e89f32cbbf32" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.967351 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tmb5s"] Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.981700 4992 scope.go:117] "RemoveContainer" containerID="916675902fc3aba5c181379f64f22d0a2a12e6528913d6eab1219844af7033f5" Oct 04 05:20:00 crc kubenswrapper[4992]: I1004 05:20:00.984893 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tmb5s"] Oct 04 05:20:01 crc kubenswrapper[4992]: E1004 05:20:01.073724 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf391f4ad_e25f_49a1_b337_3e6fc34b358d.slice/crio-ade4946362dc1e84773bb4e100cd70f985aabf756f235799fd25bb23f5c7ea4b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf391f4ad_e25f_49a1_b337_3e6fc34b358d.slice\": RecentStats: unable to find data in memory cache]" Oct 04 05:20:02 crc kubenswrapper[4992]: I1004 05:20:02.337353 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" path="/var/lib/kubelet/pods/f391f4ad-e25f-49a1-b337-3e6fc34b358d/volumes" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.855918 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w7ks5"] Oct 04 05:20:08 crc kubenswrapper[4992]: E1004 05:20:08.857870 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="registry-server" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.858917 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="registry-server" Oct 04 05:20:08 crc kubenswrapper[4992]: E1004 05:20:08.859003 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="extract-utilities" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.859057 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="extract-utilities" Oct 04 05:20:08 crc kubenswrapper[4992]: E1004 05:20:08.859267 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="extract-content" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.859333 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="extract-content" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.859748 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="f391f4ad-e25f-49a1-b337-3e6fc34b358d" containerName="registry-server" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.861727 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.870477 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w7ks5"] Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.921443 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-catalog-content\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.921491 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-utilities\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:08 crc kubenswrapper[4992]: I1004 05:20:08.921594 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq7ht\" (UniqueName: \"kubernetes.io/projected/b920d2f4-6bf1-4617-9867-a2918c261b47-kube-api-access-hq7ht\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.022776 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq7ht\" (UniqueName: \"kubernetes.io/projected/b920d2f4-6bf1-4617-9867-a2918c261b47-kube-api-access-hq7ht\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.022889 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-catalog-content\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.022915 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-utilities\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.023431 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-utilities\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.023504 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-catalog-content\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.051282 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq7ht\" (UniqueName: \"kubernetes.io/projected/b920d2f4-6bf1-4617-9867-a2918c261b47-kube-api-access-hq7ht\") pod \"certified-operators-w7ks5\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.056165 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-krbbj"] Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.058200 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.073917 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-krbbj"] Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.125009 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzfjz\" (UniqueName: \"kubernetes.io/projected/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-kube-api-access-zzfjz\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.125077 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-utilities\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.125109 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-catalog-content\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.181751 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.227186 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-utilities\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.227576 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-catalog-content\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.227789 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzfjz\" (UniqueName: \"kubernetes.io/projected/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-kube-api-access-zzfjz\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.227966 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-catalog-content\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.228174 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-utilities\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.246062 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzfjz\" (UniqueName: \"kubernetes.io/projected/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-kube-api-access-zzfjz\") pod \"community-operators-krbbj\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.410114 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.513252 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w7ks5"] Oct 04 05:20:09 crc kubenswrapper[4992]: I1004 05:20:09.970628 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-krbbj"] Oct 04 05:20:09 crc kubenswrapper[4992]: W1004 05:20:09.974200 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64b5e3cc_dd80_4e82_b425_59c6e6adf64a.slice/crio-3a1b11f534e79232e6bf8b4f9c5cc6711ad81f72cffa72a07fb89c9687e519d0 WatchSource:0}: Error finding container 3a1b11f534e79232e6bf8b4f9c5cc6711ad81f72cffa72a07fb89c9687e519d0: Status 404 returned error can't find the container with id 3a1b11f534e79232e6bf8b4f9c5cc6711ad81f72cffa72a07fb89c9687e519d0 Oct 04 05:20:10 crc kubenswrapper[4992]: I1004 05:20:10.019535 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-krbbj" event={"ID":"64b5e3cc-dd80-4e82-b425-59c6e6adf64a","Type":"ContainerStarted","Data":"3a1b11f534e79232e6bf8b4f9c5cc6711ad81f72cffa72a07fb89c9687e519d0"} Oct 04 05:20:10 crc kubenswrapper[4992]: I1004 05:20:10.020898 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7ks5" event={"ID":"b920d2f4-6bf1-4617-9867-a2918c261b47","Type":"ContainerStarted","Data":"b59419f0d31a90bdf3f476ba1e4436b61df827f14ab652a84cba4f8556b18f34"} Oct 04 05:20:11 crc kubenswrapper[4992]: I1004 05:20:11.030895 4992 generic.go:334] "Generic (PLEG): container finished" podID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerID="36a1040a82abfe6fa2ec0721cb4568b6782565175e8c28329a2ad61a287d3f21" exitCode=0 Oct 04 05:20:11 crc kubenswrapper[4992]: I1004 05:20:11.031257 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-krbbj" event={"ID":"64b5e3cc-dd80-4e82-b425-59c6e6adf64a","Type":"ContainerDied","Data":"36a1040a82abfe6fa2ec0721cb4568b6782565175e8c28329a2ad61a287d3f21"} Oct 04 05:20:11 crc kubenswrapper[4992]: I1004 05:20:11.037603 4992 generic.go:334] "Generic (PLEG): container finished" podID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerID="4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd" exitCode=0 Oct 04 05:20:11 crc kubenswrapper[4992]: I1004 05:20:11.037646 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7ks5" event={"ID":"b920d2f4-6bf1-4617-9867-a2918c261b47","Type":"ContainerDied","Data":"4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd"} Oct 04 05:20:16 crc kubenswrapper[4992]: I1004 05:20:16.086426 4992 generic.go:334] "Generic (PLEG): container finished" podID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerID="f7d950db8d8993b30727054d0e53562133626cc3e4c6149eacb41405c133fef0" exitCode=0 Oct 04 05:20:16 crc kubenswrapper[4992]: I1004 05:20:16.086526 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-krbbj" event={"ID":"64b5e3cc-dd80-4e82-b425-59c6e6adf64a","Type":"ContainerDied","Data":"f7d950db8d8993b30727054d0e53562133626cc3e4c6149eacb41405c133fef0"} Oct 04 05:20:16 crc kubenswrapper[4992]: I1004 05:20:16.089937 4992 generic.go:334] "Generic (PLEG): container finished" podID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerID="5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57" exitCode=0 Oct 04 05:20:16 crc kubenswrapper[4992]: I1004 05:20:16.089986 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7ks5" event={"ID":"b920d2f4-6bf1-4617-9867-a2918c261b47","Type":"ContainerDied","Data":"5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57"} Oct 04 05:20:17 crc kubenswrapper[4992]: I1004 05:20:17.101689 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-krbbj" event={"ID":"64b5e3cc-dd80-4e82-b425-59c6e6adf64a","Type":"ContainerStarted","Data":"af7c18f0c5235d7fd7f8602994cc421803a386b140b9b98a37de97fdf5b9fd3e"} Oct 04 05:20:18 crc kubenswrapper[4992]: I1004 05:20:18.110245 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7ks5" event={"ID":"b920d2f4-6bf1-4617-9867-a2918c261b47","Type":"ContainerStarted","Data":"587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84"} Oct 04 05:20:18 crc kubenswrapper[4992]: I1004 05:20:18.135487 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-krbbj" podStartSLOduration=3.313795463 podStartE2EDuration="9.135467395s" podCreationTimestamp="2025-10-04 05:20:09 +0000 UTC" firstStartedPulling="2025-10-04 05:20:11.034487054 +0000 UTC m=+2844.882162542" lastFinishedPulling="2025-10-04 05:20:16.856158996 +0000 UTC m=+2850.703834474" observedRunningTime="2025-10-04 05:20:17.140699315 +0000 UTC m=+2850.988374783" watchObservedRunningTime="2025-10-04 05:20:18.135467395 +0000 UTC m=+2851.983142873" Oct 04 05:20:18 crc kubenswrapper[4992]: I1004 05:20:18.138718 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w7ks5" podStartSLOduration=3.661571853 podStartE2EDuration="10.138705092s" podCreationTimestamp="2025-10-04 05:20:08 +0000 UTC" firstStartedPulling="2025-10-04 05:20:11.039548229 +0000 UTC m=+2844.887223697" lastFinishedPulling="2025-10-04 05:20:17.516681468 +0000 UTC m=+2851.364356936" observedRunningTime="2025-10-04 05:20:18.126108595 +0000 UTC m=+2851.973784063" watchObservedRunningTime="2025-10-04 05:20:18.138705092 +0000 UTC m=+2851.986380570" Oct 04 05:20:19 crc kubenswrapper[4992]: I1004 05:20:19.182457 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:19 crc kubenswrapper[4992]: I1004 05:20:19.182759 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:19 crc kubenswrapper[4992]: I1004 05:20:19.224368 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:19 crc kubenswrapper[4992]: I1004 05:20:19.411466 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:19 crc kubenswrapper[4992]: I1004 05:20:19.411529 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:19 crc kubenswrapper[4992]: I1004 05:20:19.464097 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:29 crc kubenswrapper[4992]: I1004 05:20:29.236933 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:29 crc kubenswrapper[4992]: I1004 05:20:29.311690 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w7ks5"] Oct 04 05:20:29 crc kubenswrapper[4992]: I1004 05:20:29.460568 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.215758 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w7ks5" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerName="registry-server" containerID="cri-o://587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84" gracePeriod=2 Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.800464 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.876166 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-krbbj"] Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.876754 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-krbbj" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerName="registry-server" containerID="cri-o://af7c18f0c5235d7fd7f8602994cc421803a386b140b9b98a37de97fdf5b9fd3e" gracePeriod=2 Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.913795 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-utilities\") pod \"b920d2f4-6bf1-4617-9867-a2918c261b47\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.913970 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq7ht\" (UniqueName: \"kubernetes.io/projected/b920d2f4-6bf1-4617-9867-a2918c261b47-kube-api-access-hq7ht\") pod \"b920d2f4-6bf1-4617-9867-a2918c261b47\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.913995 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-catalog-content\") pod \"b920d2f4-6bf1-4617-9867-a2918c261b47\" (UID: \"b920d2f4-6bf1-4617-9867-a2918c261b47\") " Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.915146 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-utilities" (OuterVolumeSpecName: "utilities") pod "b920d2f4-6bf1-4617-9867-a2918c261b47" (UID: "b920d2f4-6bf1-4617-9867-a2918c261b47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.920916 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b920d2f4-6bf1-4617-9867-a2918c261b47-kube-api-access-hq7ht" (OuterVolumeSpecName: "kube-api-access-hq7ht") pod "b920d2f4-6bf1-4617-9867-a2918c261b47" (UID: "b920d2f4-6bf1-4617-9867-a2918c261b47"). InnerVolumeSpecName "kube-api-access-hq7ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:20:30 crc kubenswrapper[4992]: I1004 05:20:30.958011 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b920d2f4-6bf1-4617-9867-a2918c261b47" (UID: "b920d2f4-6bf1-4617-9867-a2918c261b47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.016968 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq7ht\" (UniqueName: \"kubernetes.io/projected/b920d2f4-6bf1-4617-9867-a2918c261b47-kube-api-access-hq7ht\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.017014 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.017024 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b920d2f4-6bf1-4617-9867-a2918c261b47-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.228544 4992 generic.go:334] "Generic (PLEG): container finished" podID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerID="587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84" exitCode=0 Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.228629 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w7ks5" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.228671 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7ks5" event={"ID":"b920d2f4-6bf1-4617-9867-a2918c261b47","Type":"ContainerDied","Data":"587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84"} Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.228764 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w7ks5" event={"ID":"b920d2f4-6bf1-4617-9867-a2918c261b47","Type":"ContainerDied","Data":"b59419f0d31a90bdf3f476ba1e4436b61df827f14ab652a84cba4f8556b18f34"} Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.228795 4992 scope.go:117] "RemoveContainer" containerID="587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.232770 4992 generic.go:334] "Generic (PLEG): container finished" podID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerID="af7c18f0c5235d7fd7f8602994cc421803a386b140b9b98a37de97fdf5b9fd3e" exitCode=0 Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.232819 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-krbbj" event={"ID":"64b5e3cc-dd80-4e82-b425-59c6e6adf64a","Type":"ContainerDied","Data":"af7c18f0c5235d7fd7f8602994cc421803a386b140b9b98a37de97fdf5b9fd3e"} Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.263643 4992 scope.go:117] "RemoveContainer" containerID="5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.267631 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w7ks5"] Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.275777 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w7ks5"] Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.289177 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.298346 4992 scope.go:117] "RemoveContainer" containerID="4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.333718 4992 scope.go:117] "RemoveContainer" containerID="587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84" Oct 04 05:20:31 crc kubenswrapper[4992]: E1004 05:20:31.334471 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84\": container with ID starting with 587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84 not found: ID does not exist" containerID="587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.334515 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84"} err="failed to get container status \"587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84\": rpc error: code = NotFound desc = could not find container \"587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84\": container with ID starting with 587b4f8b3881e75a781ad47cad6d2f4d833cda014e719bcc3aa7bbcf2a186c84 not found: ID does not exist" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.334687 4992 scope.go:117] "RemoveContainer" containerID="5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57" Oct 04 05:20:31 crc kubenswrapper[4992]: E1004 05:20:31.335243 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57\": container with ID starting with 5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57 not found: ID does not exist" containerID="5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.335283 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57"} err="failed to get container status \"5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57\": rpc error: code = NotFound desc = could not find container \"5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57\": container with ID starting with 5ba6686aba68b7a4c7902102a067d9734226b705ad189939aaf45f8b4e39bb57 not found: ID does not exist" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.335305 4992 scope.go:117] "RemoveContainer" containerID="4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd" Oct 04 05:20:31 crc kubenswrapper[4992]: E1004 05:20:31.335573 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd\": container with ID starting with 4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd not found: ID does not exist" containerID="4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.335599 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd"} err="failed to get container status \"4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd\": rpc error: code = NotFound desc = could not find container \"4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd\": container with ID starting with 4915bc534518f335c2f70dd6fe9689d69699ff3546bbcd4311895ca4f232ebfd not found: ID does not exist" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.423832 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-catalog-content\") pod \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.424119 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-utilities\") pod \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.424310 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzfjz\" (UniqueName: \"kubernetes.io/projected/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-kube-api-access-zzfjz\") pod \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\" (UID: \"64b5e3cc-dd80-4e82-b425-59c6e6adf64a\") " Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.424958 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-utilities" (OuterVolumeSpecName: "utilities") pod "64b5e3cc-dd80-4e82-b425-59c6e6adf64a" (UID: "64b5e3cc-dd80-4e82-b425-59c6e6adf64a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.425495 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.428914 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-kube-api-access-zzfjz" (OuterVolumeSpecName: "kube-api-access-zzfjz") pod "64b5e3cc-dd80-4e82-b425-59c6e6adf64a" (UID: "64b5e3cc-dd80-4e82-b425-59c6e6adf64a"). InnerVolumeSpecName "kube-api-access-zzfjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.486990 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64b5e3cc-dd80-4e82-b425-59c6e6adf64a" (UID: "64b5e3cc-dd80-4e82-b425-59c6e6adf64a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.527448 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:31 crc kubenswrapper[4992]: I1004 05:20:31.527489 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzfjz\" (UniqueName: \"kubernetes.io/projected/64b5e3cc-dd80-4e82-b425-59c6e6adf64a-kube-api-access-zzfjz\") on node \"crc\" DevicePath \"\"" Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.244214 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-krbbj" event={"ID":"64b5e3cc-dd80-4e82-b425-59c6e6adf64a","Type":"ContainerDied","Data":"3a1b11f534e79232e6bf8b4f9c5cc6711ad81f72cffa72a07fb89c9687e519d0"} Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.244231 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-krbbj" Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.244278 4992 scope.go:117] "RemoveContainer" containerID="af7c18f0c5235d7fd7f8602994cc421803a386b140b9b98a37de97fdf5b9fd3e" Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.268408 4992 scope.go:117] "RemoveContainer" containerID="f7d950db8d8993b30727054d0e53562133626cc3e4c6149eacb41405c133fef0" Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.285265 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-krbbj"] Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.294751 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-krbbj"] Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.306645 4992 scope.go:117] "RemoveContainer" containerID="36a1040a82abfe6fa2ec0721cb4568b6782565175e8c28329a2ad61a287d3f21" Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.356676 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" path="/var/lib/kubelet/pods/64b5e3cc-dd80-4e82-b425-59c6e6adf64a/volumes" Oct 04 05:20:32 crc kubenswrapper[4992]: I1004 05:20:32.357933 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" path="/var/lib/kubelet/pods/b920d2f4-6bf1-4617-9867-a2918c261b47/volumes" Oct 04 05:20:50 crc kubenswrapper[4992]: I1004 05:20:50.894285 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:20:50 crc kubenswrapper[4992]: I1004 05:20:50.894914 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.345236 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ddnqf"] Oct 04 05:21:05 crc kubenswrapper[4992]: E1004 05:21:05.346567 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerName="extract-utilities" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.346600 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerName="extract-utilities" Oct 04 05:21:05 crc kubenswrapper[4992]: E1004 05:21:05.346657 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerName="extract-content" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.346672 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerName="extract-content" Oct 04 05:21:05 crc kubenswrapper[4992]: E1004 05:21:05.346707 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerName="extract-utilities" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.346720 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerName="extract-utilities" Oct 04 05:21:05 crc kubenswrapper[4992]: E1004 05:21:05.346746 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerName="registry-server" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.346763 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerName="registry-server" Oct 04 05:21:05 crc kubenswrapper[4992]: E1004 05:21:05.346790 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerName="extract-content" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.346807 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerName="extract-content" Oct 04 05:21:05 crc kubenswrapper[4992]: E1004 05:21:05.346857 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerName="registry-server" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.346870 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerName="registry-server" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.347218 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="64b5e3cc-dd80-4e82-b425-59c6e6adf64a" containerName="registry-server" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.347264 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="b920d2f4-6bf1-4617-9867-a2918c261b47" containerName="registry-server" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.349963 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.354496 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddnqf"] Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.518189 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-utilities\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.518310 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dstbv\" (UniqueName: \"kubernetes.io/projected/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-kube-api-access-dstbv\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.518433 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-catalog-content\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.620584 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dstbv\" (UniqueName: \"kubernetes.io/projected/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-kube-api-access-dstbv\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.620973 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-catalog-content\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.621185 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-utilities\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.621873 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-utilities\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.622596 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-catalog-content\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.640455 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dstbv\" (UniqueName: \"kubernetes.io/projected/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-kube-api-access-dstbv\") pod \"redhat-marketplace-ddnqf\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:05 crc kubenswrapper[4992]: I1004 05:21:05.671926 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:06 crc kubenswrapper[4992]: I1004 05:21:06.134501 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddnqf"] Oct 04 05:21:06 crc kubenswrapper[4992]: I1004 05:21:06.571710 4992 generic.go:334] "Generic (PLEG): container finished" podID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerID="956cbcb2a4d6477b73cb7951def2d0599b7b72fe6d6a83fd5f630b3c3821dfee" exitCode=0 Oct 04 05:21:06 crc kubenswrapper[4992]: I1004 05:21:06.571750 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddnqf" event={"ID":"b8bb52dd-db30-4a1b-af71-fc608c88e2ca","Type":"ContainerDied","Data":"956cbcb2a4d6477b73cb7951def2d0599b7b72fe6d6a83fd5f630b3c3821dfee"} Oct 04 05:21:06 crc kubenswrapper[4992]: I1004 05:21:06.572862 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddnqf" event={"ID":"b8bb52dd-db30-4a1b-af71-fc608c88e2ca","Type":"ContainerStarted","Data":"fe73825a3cd4cc69f0eebc717e002c119f1356b8dc44a3234cbf28060cb96383"} Oct 04 05:21:06 crc kubenswrapper[4992]: I1004 05:21:06.574067 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:21:07 crc kubenswrapper[4992]: I1004 05:21:07.582729 4992 generic.go:334] "Generic (PLEG): container finished" podID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerID="1aeafea4c9097918fef3feeb7ccb3c34aae25ba5dddd6ace7f30691d9fac930c" exitCode=0 Oct 04 05:21:07 crc kubenswrapper[4992]: I1004 05:21:07.582925 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddnqf" event={"ID":"b8bb52dd-db30-4a1b-af71-fc608c88e2ca","Type":"ContainerDied","Data":"1aeafea4c9097918fef3feeb7ccb3c34aae25ba5dddd6ace7f30691d9fac930c"} Oct 04 05:21:10 crc kubenswrapper[4992]: I1004 05:21:10.616536 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddnqf" event={"ID":"b8bb52dd-db30-4a1b-af71-fc608c88e2ca","Type":"ContainerStarted","Data":"f0919980ec3788bc2a2521f9294a1ead758f5fb02da574ccbc7c9a96e8792888"} Oct 04 05:21:10 crc kubenswrapper[4992]: I1004 05:21:10.641291 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ddnqf" podStartSLOduration=2.546885712 podStartE2EDuration="5.64127212s" podCreationTimestamp="2025-10-04 05:21:05 +0000 UTC" firstStartedPulling="2025-10-04 05:21:06.573835112 +0000 UTC m=+2900.421510580" lastFinishedPulling="2025-10-04 05:21:09.6682215 +0000 UTC m=+2903.515896988" observedRunningTime="2025-10-04 05:21:10.631891289 +0000 UTC m=+2904.479566767" watchObservedRunningTime="2025-10-04 05:21:10.64127212 +0000 UTC m=+2904.488947588" Oct 04 05:21:11 crc kubenswrapper[4992]: I1004 05:21:11.630805 4992 generic.go:334] "Generic (PLEG): container finished" podID="8e39e2be-f116-452e-ac17-3b7337d80f5d" containerID="00cb65443ce41cfd34dc22d6ec5b3e088306e686a275f56c1ce1a835871e5fc2" exitCode=0 Oct 04 05:21:11 crc kubenswrapper[4992]: I1004 05:21:11.630900 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" event={"ID":"8e39e2be-f116-452e-ac17-3b7337d80f5d","Type":"ContainerDied","Data":"00cb65443ce41cfd34dc22d6ec5b3e088306e686a275f56c1ce1a835871e5fc2"} Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.062986 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.193729 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-1\") pod \"8e39e2be-f116-452e-ac17-3b7337d80f5d\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.193797 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-inventory\") pod \"8e39e2be-f116-452e-ac17-3b7337d80f5d\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.193839 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-2\") pod \"8e39e2be-f116-452e-ac17-3b7337d80f5d\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.193945 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ssh-key\") pod \"8e39e2be-f116-452e-ac17-3b7337d80f5d\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.193983 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-528nz\" (UniqueName: \"kubernetes.io/projected/8e39e2be-f116-452e-ac17-3b7337d80f5d-kube-api-access-528nz\") pod \"8e39e2be-f116-452e-ac17-3b7337d80f5d\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.194043 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-0\") pod \"8e39e2be-f116-452e-ac17-3b7337d80f5d\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.194101 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-telemetry-combined-ca-bundle\") pod \"8e39e2be-f116-452e-ac17-3b7337d80f5d\" (UID: \"8e39e2be-f116-452e-ac17-3b7337d80f5d\") " Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.200276 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "8e39e2be-f116-452e-ac17-3b7337d80f5d" (UID: "8e39e2be-f116-452e-ac17-3b7337d80f5d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.210653 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e39e2be-f116-452e-ac17-3b7337d80f5d-kube-api-access-528nz" (OuterVolumeSpecName: "kube-api-access-528nz") pod "8e39e2be-f116-452e-ac17-3b7337d80f5d" (UID: "8e39e2be-f116-452e-ac17-3b7337d80f5d"). InnerVolumeSpecName "kube-api-access-528nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.221700 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8e39e2be-f116-452e-ac17-3b7337d80f5d" (UID: "8e39e2be-f116-452e-ac17-3b7337d80f5d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.222086 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "8e39e2be-f116-452e-ac17-3b7337d80f5d" (UID: "8e39e2be-f116-452e-ac17-3b7337d80f5d"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.227102 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-inventory" (OuterVolumeSpecName: "inventory") pod "8e39e2be-f116-452e-ac17-3b7337d80f5d" (UID: "8e39e2be-f116-452e-ac17-3b7337d80f5d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.228404 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "8e39e2be-f116-452e-ac17-3b7337d80f5d" (UID: "8e39e2be-f116-452e-ac17-3b7337d80f5d"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.232913 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "8e39e2be-f116-452e-ac17-3b7337d80f5d" (UID: "8e39e2be-f116-452e-ac17-3b7337d80f5d"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.296064 4992 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.296105 4992 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.296120 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.296136 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-528nz\" (UniqueName: \"kubernetes.io/projected/8e39e2be-f116-452e-ac17-3b7337d80f5d-kube-api-access-528nz\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.296149 4992 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.296163 4992 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.296175 4992 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8e39e2be-f116-452e-ac17-3b7337d80f5d-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.655598 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" event={"ID":"8e39e2be-f116-452e-ac17-3b7337d80f5d","Type":"ContainerDied","Data":"55ef96937f88b003c31e3faa6153bccc97ac2ac490c0190329481617e8cfa662"} Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.655914 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55ef96937f88b003c31e3faa6153bccc97ac2ac490c0190329481617e8cfa662" Oct 04 05:21:13 crc kubenswrapper[4992]: I1004 05:21:13.655671 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs" Oct 04 05:21:15 crc kubenswrapper[4992]: I1004 05:21:15.672453 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:15 crc kubenswrapper[4992]: I1004 05:21:15.672932 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:15 crc kubenswrapper[4992]: I1004 05:21:15.771362 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:16 crc kubenswrapper[4992]: I1004 05:21:16.758996 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:16 crc kubenswrapper[4992]: I1004 05:21:16.833007 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddnqf"] Oct 04 05:21:18 crc kubenswrapper[4992]: I1004 05:21:18.709089 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ddnqf" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerName="registry-server" containerID="cri-o://f0919980ec3788bc2a2521f9294a1ead758f5fb02da574ccbc7c9a96e8792888" gracePeriod=2 Oct 04 05:21:20 crc kubenswrapper[4992]: I1004 05:21:20.894263 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:21:20 crc kubenswrapper[4992]: I1004 05:21:20.894701 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:21:23 crc kubenswrapper[4992]: E1004 05:21:23.184524 4992 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8bb52dd_db30_4a1b_af71_fc608c88e2ca.slice/crio-conmon-f0919980ec3788bc2a2521f9294a1ead758f5fb02da574ccbc7c9a96e8792888.scope\": RecentStats: unable to find data in memory cache]" Oct 04 05:21:23 crc kubenswrapper[4992]: I1004 05:21:23.766927 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ddnqf_b8bb52dd-db30-4a1b-af71-fc608c88e2ca/registry-server/0.log" Oct 04 05:21:23 crc kubenswrapper[4992]: I1004 05:21:23.768124 4992 generic.go:334] "Generic (PLEG): container finished" podID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerID="f0919980ec3788bc2a2521f9294a1ead758f5fb02da574ccbc7c9a96e8792888" exitCode=137 Oct 04 05:21:23 crc kubenswrapper[4992]: I1004 05:21:23.768195 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddnqf" event={"ID":"b8bb52dd-db30-4a1b-af71-fc608c88e2ca","Type":"ContainerDied","Data":"f0919980ec3788bc2a2521f9294a1ead758f5fb02da574ccbc7c9a96e8792888"} Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.347742 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ddnqf_b8bb52dd-db30-4a1b-af71-fc608c88e2ca/registry-server/0.log" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.349860 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.441252 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dstbv\" (UniqueName: \"kubernetes.io/projected/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-kube-api-access-dstbv\") pod \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.441614 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-utilities\") pod \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.441653 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-catalog-content\") pod \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\" (UID: \"b8bb52dd-db30-4a1b-af71-fc608c88e2ca\") " Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.442622 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-utilities" (OuterVolumeSpecName: "utilities") pod "b8bb52dd-db30-4a1b-af71-fc608c88e2ca" (UID: "b8bb52dd-db30-4a1b-af71-fc608c88e2ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.447389 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-kube-api-access-dstbv" (OuterVolumeSpecName: "kube-api-access-dstbv") pod "b8bb52dd-db30-4a1b-af71-fc608c88e2ca" (UID: "b8bb52dd-db30-4a1b-af71-fc608c88e2ca"). InnerVolumeSpecName "kube-api-access-dstbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.464898 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8bb52dd-db30-4a1b-af71-fc608c88e2ca" (UID: "b8bb52dd-db30-4a1b-af71-fc608c88e2ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.544322 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.544354 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.544382 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dstbv\" (UniqueName: \"kubernetes.io/projected/b8bb52dd-db30-4a1b-af71-fc608c88e2ca-kube-api-access-dstbv\") on node \"crc\" DevicePath \"\"" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.778802 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-ddnqf_b8bb52dd-db30-4a1b-af71-fc608c88e2ca/registry-server/0.log" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.779760 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ddnqf" event={"ID":"b8bb52dd-db30-4a1b-af71-fc608c88e2ca","Type":"ContainerDied","Data":"fe73825a3cd4cc69f0eebc717e002c119f1356b8dc44a3234cbf28060cb96383"} Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.779885 4992 scope.go:117] "RemoveContainer" containerID="f0919980ec3788bc2a2521f9294a1ead758f5fb02da574ccbc7c9a96e8792888" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.779984 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ddnqf" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.809686 4992 scope.go:117] "RemoveContainer" containerID="1aeafea4c9097918fef3feeb7ccb3c34aae25ba5dddd6ace7f30691d9fac930c" Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.813177 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddnqf"] Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.822594 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ddnqf"] Oct 04 05:21:24 crc kubenswrapper[4992]: I1004 05:21:24.830039 4992 scope.go:117] "RemoveContainer" containerID="956cbcb2a4d6477b73cb7951def2d0599b7b72fe6d6a83fd5f630b3c3821dfee" Oct 04 05:21:26 crc kubenswrapper[4992]: I1004 05:21:26.334435 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" path="/var/lib/kubelet/pods/b8bb52dd-db30-4a1b-af71-fc608c88e2ca/volumes" Oct 04 05:21:50 crc kubenswrapper[4992]: I1004 05:21:50.894463 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:21:50 crc kubenswrapper[4992]: I1004 05:21:50.895051 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:21:50 crc kubenswrapper[4992]: I1004 05:21:50.895115 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 05:21:50 crc kubenswrapper[4992]: I1004 05:21:50.896176 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46df65f6a9ec684c862b0e61ea079465d75bc9be637895142632bb33563718a4"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:21:50 crc kubenswrapper[4992]: I1004 05:21:50.896274 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://46df65f6a9ec684c862b0e61ea079465d75bc9be637895142632bb33563718a4" gracePeriod=600 Oct 04 05:21:52 crc kubenswrapper[4992]: I1004 05:21:52.043887 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="46df65f6a9ec684c862b0e61ea079465d75bc9be637895142632bb33563718a4" exitCode=0 Oct 04 05:21:52 crc kubenswrapper[4992]: I1004 05:21:52.044044 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"46df65f6a9ec684c862b0e61ea079465d75bc9be637895142632bb33563718a4"} Oct 04 05:21:52 crc kubenswrapper[4992]: I1004 05:21:52.044709 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514"} Oct 04 05:21:52 crc kubenswrapper[4992]: I1004 05:21:52.044748 4992 scope.go:117] "RemoveContainer" containerID="c5b362d76fa02aeb39d73973e735a51b819dc070cc16da47e7cdbf8bab200505" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.019406 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 05:22:16 crc kubenswrapper[4992]: E1004 05:22:16.020347 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerName="extract-utilities" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.020379 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerName="extract-utilities" Oct 04 05:22:16 crc kubenswrapper[4992]: E1004 05:22:16.020404 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e39e2be-f116-452e-ac17-3b7337d80f5d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.020414 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e39e2be-f116-452e-ac17-3b7337d80f5d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 05:22:16 crc kubenswrapper[4992]: E1004 05:22:16.020436 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerName="extract-content" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.020444 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerName="extract-content" Oct 04 05:22:16 crc kubenswrapper[4992]: E1004 05:22:16.020455 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerName="registry-server" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.020463 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerName="registry-server" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.020700 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e39e2be-f116-452e-ac17-3b7337d80f5d" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.020733 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8bb52dd-db30-4a1b-af71-fc608c88e2ca" containerName="registry-server" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.021546 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.023384 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.023699 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.025132 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.030919 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pr8fk" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.036696 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.156761 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.156897 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.156982 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.157019 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.157132 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt9bw\" (UniqueName: \"kubernetes.io/projected/dab26aab-5877-4a73-9433-7f76c31377dc-kube-api-access-mt9bw\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.157188 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.157211 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.157277 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.157375 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-config-data\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259166 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-config-data\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259300 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259338 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259388 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259420 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259448 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt9bw\" (UniqueName: \"kubernetes.io/projected/dab26aab-5877-4a73-9433-7f76c31377dc-kube-api-access-mt9bw\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259476 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259496 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.259568 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.260225 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.260258 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.260735 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.261059 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.261065 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-config-data\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.281811 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.283799 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.284325 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.284983 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt9bw\" (UniqueName: \"kubernetes.io/projected/dab26aab-5877-4a73-9433-7f76c31377dc-kube-api-access-mt9bw\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.304535 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.339916 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 05:22:16 crc kubenswrapper[4992]: I1004 05:22:16.777748 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 05:22:17 crc kubenswrapper[4992]: I1004 05:22:17.285916 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"dab26aab-5877-4a73-9433-7f76c31377dc","Type":"ContainerStarted","Data":"f3491616396043d01cf8d840d8e108a0f756c58357a2dc20456013681c6b93aa"} Oct 04 05:23:11 crc kubenswrapper[4992]: E1004 05:23:11.005247 4992 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 04 05:23:11 crc kubenswrapper[4992]: E1004 05:23:11.006023 4992 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mt9bw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(dab26aab-5877-4a73-9433-7f76c31377dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 05:23:11 crc kubenswrapper[4992]: E1004 05:23:11.007287 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="dab26aab-5877-4a73-9433-7f76c31377dc" Oct 04 05:23:11 crc kubenswrapper[4992]: E1004 05:23:11.826354 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="dab26aab-5877-4a73-9433-7f76c31377dc" Oct 04 05:23:26 crc kubenswrapper[4992]: I1004 05:23:26.823952 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 05:23:27 crc kubenswrapper[4992]: I1004 05:23:27.980449 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"dab26aab-5877-4a73-9433-7f76c31377dc","Type":"ContainerStarted","Data":"e4780c955e26f441d9867a104d5e2368386bdbcd4837f20611f06b2d576a0964"} Oct 04 05:23:28 crc kubenswrapper[4992]: I1004 05:23:28.000432 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.9739839679999998 podStartE2EDuration="1m14.000416742s" podCreationTimestamp="2025-10-04 05:22:14 +0000 UTC" firstStartedPulling="2025-10-04 05:22:16.794875285 +0000 UTC m=+2970.642550753" lastFinishedPulling="2025-10-04 05:23:26.821308069 +0000 UTC m=+3040.668983527" observedRunningTime="2025-10-04 05:23:27.997458713 +0000 UTC m=+3041.845134191" watchObservedRunningTime="2025-10-04 05:23:28.000416742 +0000 UTC m=+3041.848092210" Oct 04 05:24:20 crc kubenswrapper[4992]: I1004 05:24:20.894700 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:24:20 crc kubenswrapper[4992]: I1004 05:24:20.895229 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:24:50 crc kubenswrapper[4992]: I1004 05:24:50.894467 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:24:50 crc kubenswrapper[4992]: I1004 05:24:50.895069 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:25:20 crc kubenswrapper[4992]: I1004 05:25:20.895118 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:25:20 crc kubenswrapper[4992]: I1004 05:25:20.895854 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:25:20 crc kubenswrapper[4992]: I1004 05:25:20.895921 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 05:25:20 crc kubenswrapper[4992]: I1004 05:25:20.896862 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:25:20 crc kubenswrapper[4992]: I1004 05:25:20.896969 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" gracePeriod=600 Oct 04 05:25:21 crc kubenswrapper[4992]: E1004 05:25:21.067953 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:25:21 crc kubenswrapper[4992]: I1004 05:25:21.096052 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" exitCode=0 Oct 04 05:25:21 crc kubenswrapper[4992]: I1004 05:25:21.096094 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514"} Oct 04 05:25:21 crc kubenswrapper[4992]: I1004 05:25:21.096124 4992 scope.go:117] "RemoveContainer" containerID="46df65f6a9ec684c862b0e61ea079465d75bc9be637895142632bb33563718a4" Oct 04 05:25:21 crc kubenswrapper[4992]: I1004 05:25:21.099778 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:25:21 crc kubenswrapper[4992]: E1004 05:25:21.100617 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:25:34 crc kubenswrapper[4992]: I1004 05:25:34.341209 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:25:34 crc kubenswrapper[4992]: E1004 05:25:34.342524 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:25:46 crc kubenswrapper[4992]: I1004 05:25:46.319685 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:25:46 crc kubenswrapper[4992]: E1004 05:25:46.320432 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:26:01 crc kubenswrapper[4992]: I1004 05:26:01.319733 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:26:01 crc kubenswrapper[4992]: E1004 05:26:01.320521 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:26:13 crc kubenswrapper[4992]: I1004 05:26:13.320494 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:26:13 crc kubenswrapper[4992]: E1004 05:26:13.321279 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:26:25 crc kubenswrapper[4992]: I1004 05:26:25.319793 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:26:25 crc kubenswrapper[4992]: E1004 05:26:25.320589 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:26:36 crc kubenswrapper[4992]: I1004 05:26:36.320088 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:26:36 crc kubenswrapper[4992]: E1004 05:26:36.321150 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:26:50 crc kubenswrapper[4992]: I1004 05:26:50.319408 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:26:50 crc kubenswrapper[4992]: E1004 05:26:50.320127 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:27:05 crc kubenswrapper[4992]: I1004 05:27:05.319257 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:27:05 crc kubenswrapper[4992]: E1004 05:27:05.320202 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:27:19 crc kubenswrapper[4992]: I1004 05:27:19.319376 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:27:19 crc kubenswrapper[4992]: E1004 05:27:19.320280 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:27:33 crc kubenswrapper[4992]: I1004 05:27:33.319643 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:27:33 crc kubenswrapper[4992]: E1004 05:27:33.321282 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:27:47 crc kubenswrapper[4992]: I1004 05:27:47.319410 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:27:47 crc kubenswrapper[4992]: E1004 05:27:47.320393 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:27:59 crc kubenswrapper[4992]: I1004 05:27:59.319401 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:27:59 crc kubenswrapper[4992]: E1004 05:27:59.320214 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:28:13 crc kubenswrapper[4992]: I1004 05:28:13.318883 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:28:13 crc kubenswrapper[4992]: E1004 05:28:13.319720 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:28:28 crc kubenswrapper[4992]: I1004 05:28:28.319409 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:28:28 crc kubenswrapper[4992]: E1004 05:28:28.320264 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:28:41 crc kubenswrapper[4992]: I1004 05:28:41.319514 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:28:41 crc kubenswrapper[4992]: E1004 05:28:41.320150 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:28:53 crc kubenswrapper[4992]: I1004 05:28:53.318651 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:28:53 crc kubenswrapper[4992]: E1004 05:28:53.319402 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:29:04 crc kubenswrapper[4992]: I1004 05:29:04.327908 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:29:04 crc kubenswrapper[4992]: E1004 05:29:04.328818 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:29:16 crc kubenswrapper[4992]: I1004 05:29:16.320030 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:29:16 crc kubenswrapper[4992]: E1004 05:29:16.321182 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:29:28 crc kubenswrapper[4992]: I1004 05:29:28.356284 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:29:28 crc kubenswrapper[4992]: E1004 05:29:28.357355 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:29:43 crc kubenswrapper[4992]: I1004 05:29:43.320055 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:29:43 crc kubenswrapper[4992]: E1004 05:29:43.322660 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:29:55 crc kubenswrapper[4992]: I1004 05:29:55.320432 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:29:55 crc kubenswrapper[4992]: E1004 05:29:55.322345 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.170143 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq"] Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.171868 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.173752 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.174030 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.186756 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq"] Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.311753 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6w5j\" (UniqueName: \"kubernetes.io/projected/3d520648-ce74-4296-8de1-a2d9c567482e-kube-api-access-x6w5j\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.311807 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d520648-ce74-4296-8de1-a2d9c567482e-config-volume\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.311841 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d520648-ce74-4296-8de1-a2d9c567482e-secret-volume\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.414311 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6w5j\" (UniqueName: \"kubernetes.io/projected/3d520648-ce74-4296-8de1-a2d9c567482e-kube-api-access-x6w5j\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.414446 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d520648-ce74-4296-8de1-a2d9c567482e-config-volume\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.414517 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d520648-ce74-4296-8de1-a2d9c567482e-secret-volume\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.415701 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d520648-ce74-4296-8de1-a2d9c567482e-config-volume\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.421768 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d520648-ce74-4296-8de1-a2d9c567482e-secret-volume\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.446520 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6w5j\" (UniqueName: \"kubernetes.io/projected/3d520648-ce74-4296-8de1-a2d9c567482e-kube-api-access-x6w5j\") pod \"collect-profiles-29325930-8qzqq\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.495477 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:00 crc kubenswrapper[4992]: I1004 05:30:00.938196 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq"] Oct 04 05:30:01 crc kubenswrapper[4992]: I1004 05:30:01.783441 4992 generic.go:334] "Generic (PLEG): container finished" podID="3d520648-ce74-4296-8de1-a2d9c567482e" containerID="1045bb3c0d26a52b17447bc093236143f04842de5e21544afa8b84ce2bfbab7f" exitCode=0 Oct 04 05:30:01 crc kubenswrapper[4992]: I1004 05:30:01.783510 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" event={"ID":"3d520648-ce74-4296-8de1-a2d9c567482e","Type":"ContainerDied","Data":"1045bb3c0d26a52b17447bc093236143f04842de5e21544afa8b84ce2bfbab7f"} Oct 04 05:30:01 crc kubenswrapper[4992]: I1004 05:30:01.783837 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" event={"ID":"3d520648-ce74-4296-8de1-a2d9c567482e","Type":"ContainerStarted","Data":"d0460c16e073b84eaf30442da4efd88c7448aef0714d30c6e3a19e0c0d558d5d"} Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.187380 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.281567 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6w5j\" (UniqueName: \"kubernetes.io/projected/3d520648-ce74-4296-8de1-a2d9c567482e-kube-api-access-x6w5j\") pod \"3d520648-ce74-4296-8de1-a2d9c567482e\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.281674 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d520648-ce74-4296-8de1-a2d9c567482e-secret-volume\") pod \"3d520648-ce74-4296-8de1-a2d9c567482e\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.281894 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d520648-ce74-4296-8de1-a2d9c567482e-config-volume\") pod \"3d520648-ce74-4296-8de1-a2d9c567482e\" (UID: \"3d520648-ce74-4296-8de1-a2d9c567482e\") " Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.282647 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d520648-ce74-4296-8de1-a2d9c567482e-config-volume" (OuterVolumeSpecName: "config-volume") pod "3d520648-ce74-4296-8de1-a2d9c567482e" (UID: "3d520648-ce74-4296-8de1-a2d9c567482e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.287986 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d520648-ce74-4296-8de1-a2d9c567482e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3d520648-ce74-4296-8de1-a2d9c567482e" (UID: "3d520648-ce74-4296-8de1-a2d9c567482e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.288485 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d520648-ce74-4296-8de1-a2d9c567482e-kube-api-access-x6w5j" (OuterVolumeSpecName: "kube-api-access-x6w5j") pod "3d520648-ce74-4296-8de1-a2d9c567482e" (UID: "3d520648-ce74-4296-8de1-a2d9c567482e"). InnerVolumeSpecName "kube-api-access-x6w5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.384492 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6w5j\" (UniqueName: \"kubernetes.io/projected/3d520648-ce74-4296-8de1-a2d9c567482e-kube-api-access-x6w5j\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.384873 4992 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3d520648-ce74-4296-8de1-a2d9c567482e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.384890 4992 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3d520648-ce74-4296-8de1-a2d9c567482e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.805032 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" event={"ID":"3d520648-ce74-4296-8de1-a2d9c567482e","Type":"ContainerDied","Data":"d0460c16e073b84eaf30442da4efd88c7448aef0714d30c6e3a19e0c0d558d5d"} Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.805316 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0460c16e073b84eaf30442da4efd88c7448aef0714d30c6e3a19e0c0d558d5d" Oct 04 05:30:03 crc kubenswrapper[4992]: I1004 05:30:03.805117 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325930-8qzqq" Oct 04 05:30:04 crc kubenswrapper[4992]: I1004 05:30:04.255723 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66"] Oct 04 05:30:04 crc kubenswrapper[4992]: I1004 05:30:04.269869 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325885-87l66"] Oct 04 05:30:04 crc kubenswrapper[4992]: I1004 05:30:04.332906 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5994ffa6-01dd-4b1c-b592-be6f6845f20a" path="/var/lib/kubelet/pods/5994ffa6-01dd-4b1c-b592-be6f6845f20a/volumes" Oct 04 05:30:06 crc kubenswrapper[4992]: I1004 05:30:06.319683 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:30:06 crc kubenswrapper[4992]: E1004 05:30:06.320520 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:30:12 crc kubenswrapper[4992]: I1004 05:30:12.985016 4992 scope.go:117] "RemoveContainer" containerID="2bd0d4cfad0092c0b1428d6cd51b868fb9da1bf1286b48a137db4334e46c5a1e" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.460482 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lcm7t"] Oct 04 05:30:20 crc kubenswrapper[4992]: E1004 05:30:20.462459 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d520648-ce74-4296-8de1-a2d9c567482e" containerName="collect-profiles" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.462492 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d520648-ce74-4296-8de1-a2d9c567482e" containerName="collect-profiles" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.463065 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d520648-ce74-4296-8de1-a2d9c567482e" containerName="collect-profiles" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.466491 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.476401 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lcm7t"] Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.526739 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbbm9\" (UniqueName: \"kubernetes.io/projected/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-kube-api-access-qbbm9\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.527028 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-utilities\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.527208 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-catalog-content\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.629129 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbbm9\" (UniqueName: \"kubernetes.io/projected/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-kube-api-access-qbbm9\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.629270 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-utilities\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.629350 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-catalog-content\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.629956 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-utilities\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.630016 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-catalog-content\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.652224 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbbm9\" (UniqueName: \"kubernetes.io/projected/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-kube-api-access-qbbm9\") pod \"certified-operators-lcm7t\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:20 crc kubenswrapper[4992]: I1004 05:30:20.798767 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:21 crc kubenswrapper[4992]: I1004 05:30:21.259589 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lcm7t"] Oct 04 05:30:21 crc kubenswrapper[4992]: I1004 05:30:21.319432 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:30:21 crc kubenswrapper[4992]: I1004 05:30:21.953538 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcm7t" event={"ID":"63cf10c7-7ce1-4cdb-becd-361b7f47ca44","Type":"ContainerStarted","Data":"f5c0542bc81e70744b28288dc6d3a5dfa4a87a25e261399ad9c7f9232299824b"} Oct 04 05:30:26 crc kubenswrapper[4992]: I1004 05:30:26.007994 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcm7t" event={"ID":"63cf10c7-7ce1-4cdb-becd-361b7f47ca44","Type":"ContainerStarted","Data":"b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832"} Oct 04 05:30:27 crc kubenswrapper[4992]: I1004 05:30:27.018197 4992 generic.go:334] "Generic (PLEG): container finished" podID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerID="b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832" exitCode=0 Oct 04 05:30:27 crc kubenswrapper[4992]: I1004 05:30:27.018316 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcm7t" event={"ID":"63cf10c7-7ce1-4cdb-becd-361b7f47ca44","Type":"ContainerDied","Data":"b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832"} Oct 04 05:30:27 crc kubenswrapper[4992]: I1004 05:30:27.022417 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"95a2327fa1567e14683aac6a2334df02751530ba6b438478d3fecaa93c2aa785"} Oct 04 05:30:28 crc kubenswrapper[4992]: I1004 05:30:28.038120 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.413957 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pn7c7"] Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.419266 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.434425 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pn7c7"] Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.536849 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzr7w\" (UniqueName: \"kubernetes.io/projected/aa70c135-0efc-4f2e-bf72-25b6055cda88-kube-api-access-dzr7w\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.536925 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa70c135-0efc-4f2e-bf72-25b6055cda88-utilities\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.537037 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa70c135-0efc-4f2e-bf72-25b6055cda88-catalog-content\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.638812 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzr7w\" (UniqueName: \"kubernetes.io/projected/aa70c135-0efc-4f2e-bf72-25b6055cda88-kube-api-access-dzr7w\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.639104 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa70c135-0efc-4f2e-bf72-25b6055cda88-utilities\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.639205 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa70c135-0efc-4f2e-bf72-25b6055cda88-catalog-content\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.639651 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa70c135-0efc-4f2e-bf72-25b6055cda88-utilities\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.639758 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa70c135-0efc-4f2e-bf72-25b6055cda88-catalog-content\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.661526 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzr7w\" (UniqueName: \"kubernetes.io/projected/aa70c135-0efc-4f2e-bf72-25b6055cda88-kube-api-access-dzr7w\") pod \"community-operators-pn7c7\" (UID: \"aa70c135-0efc-4f2e-bf72-25b6055cda88\") " pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:29 crc kubenswrapper[4992]: I1004 05:30:29.748830 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:30 crc kubenswrapper[4992]: I1004 05:30:30.957394 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pn7c7"] Oct 04 05:30:31 crc kubenswrapper[4992]: I1004 05:30:31.061949 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn7c7" event={"ID":"aa70c135-0efc-4f2e-bf72-25b6055cda88","Type":"ContainerStarted","Data":"f14dcf7c7fd534503b0188464c1be95f93bfbdb8c6c6b8f461b86e0b0d6092ef"} Oct 04 05:30:31 crc kubenswrapper[4992]: I1004 05:30:31.064955 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcm7t" event={"ID":"63cf10c7-7ce1-4cdb-becd-361b7f47ca44","Type":"ContainerStarted","Data":"c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3"} Oct 04 05:30:32 crc kubenswrapper[4992]: I1004 05:30:32.073968 4992 generic.go:334] "Generic (PLEG): container finished" podID="aa70c135-0efc-4f2e-bf72-25b6055cda88" containerID="ff10e4937664e2651ecb41cd7300f301b36fde6c60f6ffb2259f470d77d4fc95" exitCode=0 Oct 04 05:30:32 crc kubenswrapper[4992]: I1004 05:30:32.074019 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn7c7" event={"ID":"aa70c135-0efc-4f2e-bf72-25b6055cda88","Type":"ContainerDied","Data":"ff10e4937664e2651ecb41cd7300f301b36fde6c60f6ffb2259f470d77d4fc95"} Oct 04 05:30:32 crc kubenswrapper[4992]: I1004 05:30:32.076397 4992 generic.go:334] "Generic (PLEG): container finished" podID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerID="c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3" exitCode=0 Oct 04 05:30:32 crc kubenswrapper[4992]: I1004 05:30:32.076434 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcm7t" event={"ID":"63cf10c7-7ce1-4cdb-becd-361b7f47ca44","Type":"ContainerDied","Data":"c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3"} Oct 04 05:30:35 crc kubenswrapper[4992]: I1004 05:30:35.108998 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcm7t" event={"ID":"63cf10c7-7ce1-4cdb-becd-361b7f47ca44","Type":"ContainerStarted","Data":"74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb"} Oct 04 05:30:35 crc kubenswrapper[4992]: I1004 05:30:35.133026 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lcm7t" podStartSLOduration=9.088132532 podStartE2EDuration="15.132991107s" podCreationTimestamp="2025-10-04 05:30:20 +0000 UTC" firstStartedPulling="2025-10-04 05:30:28.037591543 +0000 UTC m=+3461.885267041" lastFinishedPulling="2025-10-04 05:30:34.082450148 +0000 UTC m=+3467.930125616" observedRunningTime="2025-10-04 05:30:35.126966436 +0000 UTC m=+3468.974641904" watchObservedRunningTime="2025-10-04 05:30:35.132991107 +0000 UTC m=+3468.980666575" Oct 04 05:30:40 crc kubenswrapper[4992]: I1004 05:30:40.163908 4992 generic.go:334] "Generic (PLEG): container finished" podID="aa70c135-0efc-4f2e-bf72-25b6055cda88" containerID="1c16d815ac00d5ddd36676947aed1dab8894abe6b7fff5eb20cae9b90f1b6885" exitCode=0 Oct 04 05:30:40 crc kubenswrapper[4992]: I1004 05:30:40.163991 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn7c7" event={"ID":"aa70c135-0efc-4f2e-bf72-25b6055cda88","Type":"ContainerDied","Data":"1c16d815ac00d5ddd36676947aed1dab8894abe6b7fff5eb20cae9b90f1b6885"} Oct 04 05:30:40 crc kubenswrapper[4992]: I1004 05:30:40.798942 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:40 crc kubenswrapper[4992]: I1004 05:30:40.799266 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:40 crc kubenswrapper[4992]: I1004 05:30:40.854975 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:41 crc kubenswrapper[4992]: I1004 05:30:41.177150 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pn7c7" event={"ID":"aa70c135-0efc-4f2e-bf72-25b6055cda88","Type":"ContainerStarted","Data":"884e644a3bb16d59fa1f5d0c6da1679936045cb7ddbfeb00cbd2166d2523392a"} Oct 04 05:30:41 crc kubenswrapper[4992]: I1004 05:30:41.206863 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pn7c7" podStartSLOduration=3.501360128 podStartE2EDuration="12.206813685s" podCreationTimestamp="2025-10-04 05:30:29 +0000 UTC" firstStartedPulling="2025-10-04 05:30:32.075786831 +0000 UTC m=+3465.923462299" lastFinishedPulling="2025-10-04 05:30:40.781240368 +0000 UTC m=+3474.628915856" observedRunningTime="2025-10-04 05:30:41.205028947 +0000 UTC m=+3475.052704415" watchObservedRunningTime="2025-10-04 05:30:41.206813685 +0000 UTC m=+3475.054489153" Oct 04 05:30:41 crc kubenswrapper[4992]: I1004 05:30:41.226718 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:42 crc kubenswrapper[4992]: I1004 05:30:42.802469 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lcm7t"] Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.197975 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lcm7t" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerName="registry-server" containerID="cri-o://74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb" gracePeriod=2 Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.649047 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.732242 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-catalog-content\") pod \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.732323 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbbm9\" (UniqueName: \"kubernetes.io/projected/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-kube-api-access-qbbm9\") pod \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.732426 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-utilities\") pod \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\" (UID: \"63cf10c7-7ce1-4cdb-becd-361b7f47ca44\") " Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.733049 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-utilities" (OuterVolumeSpecName: "utilities") pod "63cf10c7-7ce1-4cdb-becd-361b7f47ca44" (UID: "63cf10c7-7ce1-4cdb-becd-361b7f47ca44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.738439 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-kube-api-access-qbbm9" (OuterVolumeSpecName: "kube-api-access-qbbm9") pod "63cf10c7-7ce1-4cdb-becd-361b7f47ca44" (UID: "63cf10c7-7ce1-4cdb-becd-361b7f47ca44"). InnerVolumeSpecName "kube-api-access-qbbm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.779529 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63cf10c7-7ce1-4cdb-becd-361b7f47ca44" (UID: "63cf10c7-7ce1-4cdb-becd-361b7f47ca44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.835034 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.835082 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:43 crc kubenswrapper[4992]: I1004 05:30:43.835099 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbbm9\" (UniqueName: \"kubernetes.io/projected/63cf10c7-7ce1-4cdb-becd-361b7f47ca44-kube-api-access-qbbm9\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.211033 4992 generic.go:334] "Generic (PLEG): container finished" podID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerID="74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb" exitCode=0 Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.211075 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcm7t" event={"ID":"63cf10c7-7ce1-4cdb-becd-361b7f47ca44","Type":"ContainerDied","Data":"74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb"} Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.211101 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcm7t" event={"ID":"63cf10c7-7ce1-4cdb-becd-361b7f47ca44","Type":"ContainerDied","Data":"f5c0542bc81e70744b28288dc6d3a5dfa4a87a25e261399ad9c7f9232299824b"} Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.211120 4992 scope.go:117] "RemoveContainer" containerID="74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.211275 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcm7t" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.216451 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-67msj"] Oct 04 05:30:44 crc kubenswrapper[4992]: E1004 05:30:44.216923 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerName="registry-server" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.216940 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerName="registry-server" Oct 04 05:30:44 crc kubenswrapper[4992]: E1004 05:30:44.216960 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerName="extract-utilities" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.216968 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerName="extract-utilities" Oct 04 05:30:44 crc kubenswrapper[4992]: E1004 05:30:44.217005 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerName="extract-content" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.217012 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerName="extract-content" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.217222 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" containerName="registry-server" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.218698 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.224886 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67msj"] Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.257666 4992 scope.go:117] "RemoveContainer" containerID="c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.284802 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lcm7t"] Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.292485 4992 scope.go:117] "RemoveContainer" containerID="b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.303889 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lcm7t"] Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.355988 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-catalog-content\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.356592 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfgw5\" (UniqueName: \"kubernetes.io/projected/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-kube-api-access-wfgw5\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.356795 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-utilities\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.358768 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63cf10c7-7ce1-4cdb-becd-361b7f47ca44" path="/var/lib/kubelet/pods/63cf10c7-7ce1-4cdb-becd-361b7f47ca44/volumes" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.458981 4992 scope.go:117] "RemoveContainer" containerID="74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.459062 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-catalog-content\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.459352 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfgw5\" (UniqueName: \"kubernetes.io/projected/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-kube-api-access-wfgw5\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.459458 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-utilities\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.460013 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-utilities\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.460326 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-catalog-content\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: E1004 05:30:44.460468 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb\": container with ID starting with 74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb not found: ID does not exist" containerID="74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.460580 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb"} err="failed to get container status \"74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb\": rpc error: code = NotFound desc = could not find container \"74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb\": container with ID starting with 74e151307c293464e8b84414fb6c51f58a998c753fd2627366dbaffa0cf92bbb not found: ID does not exist" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.460680 4992 scope.go:117] "RemoveContainer" containerID="c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3" Oct 04 05:30:44 crc kubenswrapper[4992]: E1004 05:30:44.462850 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3\": container with ID starting with c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3 not found: ID does not exist" containerID="c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.463040 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3"} err="failed to get container status \"c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3\": rpc error: code = NotFound desc = could not find container \"c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3\": container with ID starting with c28041c520f0d09c2d14dc964fa8a741c64f77aefda333f808af1354eed915b3 not found: ID does not exist" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.463138 4992 scope.go:117] "RemoveContainer" containerID="b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832" Oct 04 05:30:44 crc kubenswrapper[4992]: E1004 05:30:44.463531 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832\": container with ID starting with b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832 not found: ID does not exist" containerID="b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.463568 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832"} err="failed to get container status \"b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832\": rpc error: code = NotFound desc = could not find container \"b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832\": container with ID starting with b7f4ccb5af4040a40dcc794fc3137e0317949af3ab4dfad4b9b44fbe0ba53832 not found: ID does not exist" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.482762 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfgw5\" (UniqueName: \"kubernetes.io/projected/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-kube-api-access-wfgw5\") pod \"redhat-operators-67msj\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:44 crc kubenswrapper[4992]: I1004 05:30:44.549256 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:45 crc kubenswrapper[4992]: W1004 05:30:45.017268 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e8f0b2e_d98a_4939_88b5_7cb82bacb692.slice/crio-485a0c18b6e9e376d52af54280772ec15a4228be40c121abe5a740e3c2d8f75a WatchSource:0}: Error finding container 485a0c18b6e9e376d52af54280772ec15a4228be40c121abe5a740e3c2d8f75a: Status 404 returned error can't find the container with id 485a0c18b6e9e376d52af54280772ec15a4228be40c121abe5a740e3c2d8f75a Oct 04 05:30:45 crc kubenswrapper[4992]: I1004 05:30:45.031674 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67msj"] Oct 04 05:30:45 crc kubenswrapper[4992]: I1004 05:30:45.221253 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67msj" event={"ID":"3e8f0b2e-d98a-4939-88b5-7cb82bacb692","Type":"ContainerStarted","Data":"485a0c18b6e9e376d52af54280772ec15a4228be40c121abe5a740e3c2d8f75a"} Oct 04 05:30:48 crc kubenswrapper[4992]: I1004 05:30:48.250235 4992 generic.go:334] "Generic (PLEG): container finished" podID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerID="1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b" exitCode=0 Oct 04 05:30:48 crc kubenswrapper[4992]: I1004 05:30:48.250398 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67msj" event={"ID":"3e8f0b2e-d98a-4939-88b5-7cb82bacb692","Type":"ContainerDied","Data":"1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b"} Oct 04 05:30:49 crc kubenswrapper[4992]: I1004 05:30:49.750558 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:49 crc kubenswrapper[4992]: I1004 05:30:49.751100 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:49 crc kubenswrapper[4992]: I1004 05:30:49.834465 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:50 crc kubenswrapper[4992]: I1004 05:30:50.387739 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pn7c7" Oct 04 05:30:50 crc kubenswrapper[4992]: I1004 05:30:50.695784 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pn7c7"] Oct 04 05:30:51 crc kubenswrapper[4992]: I1004 05:30:51.002488 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s55dk"] Oct 04 05:30:51 crc kubenswrapper[4992]: I1004 05:30:51.002728 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s55dk" podUID="e208c780-5970-45e0-adce-55980aa4bd88" containerName="registry-server" containerID="cri-o://89d646e67c891a5fb50ea641c0ae6168148cf2c34cc8cbc665bd62ce609f59c9" gracePeriod=2 Oct 04 05:30:51 crc kubenswrapper[4992]: I1004 05:30:51.347805 4992 generic.go:334] "Generic (PLEG): container finished" podID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerID="e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528" exitCode=0 Oct 04 05:30:51 crc kubenswrapper[4992]: I1004 05:30:51.347859 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67msj" event={"ID":"3e8f0b2e-d98a-4939-88b5-7cb82bacb692","Type":"ContainerDied","Data":"e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528"} Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.358655 4992 generic.go:334] "Generic (PLEG): container finished" podID="e208c780-5970-45e0-adce-55980aa4bd88" containerID="89d646e67c891a5fb50ea641c0ae6168148cf2c34cc8cbc665bd62ce609f59c9" exitCode=0 Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.358746 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s55dk" event={"ID":"e208c780-5970-45e0-adce-55980aa4bd88","Type":"ContainerDied","Data":"89d646e67c891a5fb50ea641c0ae6168148cf2c34cc8cbc665bd62ce609f59c9"} Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.359373 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s55dk" event={"ID":"e208c780-5970-45e0-adce-55980aa4bd88","Type":"ContainerDied","Data":"b9dac6345f7b1747198c4c15d0e7f3fe002a6ec4f557f4b15ef2c61c0d6037b6"} Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.359391 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9dac6345f7b1747198c4c15d0e7f3fe002a6ec4f557f4b15ef2c61c0d6037b6" Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.414256 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s55dk" Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.599379 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-catalog-content\") pod \"e208c780-5970-45e0-adce-55980aa4bd88\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.599595 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-utilities\") pod \"e208c780-5970-45e0-adce-55980aa4bd88\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.599684 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmxc2\" (UniqueName: \"kubernetes.io/projected/e208c780-5970-45e0-adce-55980aa4bd88-kube-api-access-kmxc2\") pod \"e208c780-5970-45e0-adce-55980aa4bd88\" (UID: \"e208c780-5970-45e0-adce-55980aa4bd88\") " Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.600896 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-utilities" (OuterVolumeSpecName: "utilities") pod "e208c780-5970-45e0-adce-55980aa4bd88" (UID: "e208c780-5970-45e0-adce-55980aa4bd88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.613197 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e208c780-5970-45e0-adce-55980aa4bd88-kube-api-access-kmxc2" (OuterVolumeSpecName: "kube-api-access-kmxc2") pod "e208c780-5970-45e0-adce-55980aa4bd88" (UID: "e208c780-5970-45e0-adce-55980aa4bd88"). InnerVolumeSpecName "kube-api-access-kmxc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.641380 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e208c780-5970-45e0-adce-55980aa4bd88" (UID: "e208c780-5970-45e0-adce-55980aa4bd88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.701775 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.702432 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmxc2\" (UniqueName: \"kubernetes.io/projected/e208c780-5970-45e0-adce-55980aa4bd88-kube-api-access-kmxc2\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:52 crc kubenswrapper[4992]: I1004 05:30:52.702470 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e208c780-5970-45e0-adce-55980aa4bd88-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:30:53 crc kubenswrapper[4992]: I1004 05:30:53.367029 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s55dk" Oct 04 05:30:53 crc kubenswrapper[4992]: I1004 05:30:53.400373 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s55dk"] Oct 04 05:30:53 crc kubenswrapper[4992]: I1004 05:30:53.409332 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s55dk"] Oct 04 05:30:54 crc kubenswrapper[4992]: I1004 05:30:54.338434 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e208c780-5970-45e0-adce-55980aa4bd88" path="/var/lib/kubelet/pods/e208c780-5970-45e0-adce-55980aa4bd88/volumes" Oct 04 05:30:54 crc kubenswrapper[4992]: I1004 05:30:54.379872 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67msj" event={"ID":"3e8f0b2e-d98a-4939-88b5-7cb82bacb692","Type":"ContainerStarted","Data":"710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334"} Oct 04 05:30:54 crc kubenswrapper[4992]: I1004 05:30:54.407415 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-67msj" podStartSLOduration=5.192343163 podStartE2EDuration="10.407395002s" podCreationTimestamp="2025-10-04 05:30:44 +0000 UTC" firstStartedPulling="2025-10-04 05:30:48.252097189 +0000 UTC m=+3482.099772657" lastFinishedPulling="2025-10-04 05:30:53.467149028 +0000 UTC m=+3487.314824496" observedRunningTime="2025-10-04 05:30:54.400889318 +0000 UTC m=+3488.248564826" watchObservedRunningTime="2025-10-04 05:30:54.407395002 +0000 UTC m=+3488.255070480" Oct 04 05:30:54 crc kubenswrapper[4992]: I1004 05:30:54.552328 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:54 crc kubenswrapper[4992]: I1004 05:30:54.552391 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:30:55 crc kubenswrapper[4992]: I1004 05:30:55.595288 4992 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-67msj" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="registry-server" probeResult="failure" output=< Oct 04 05:30:55 crc kubenswrapper[4992]: timeout: failed to connect service ":50051" within 1s Oct 04 05:30:55 crc kubenswrapper[4992]: > Oct 04 05:31:04 crc kubenswrapper[4992]: I1004 05:31:04.602475 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:31:04 crc kubenswrapper[4992]: I1004 05:31:04.649938 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:31:04 crc kubenswrapper[4992]: I1004 05:31:04.838095 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67msj"] Oct 04 05:31:06 crc kubenswrapper[4992]: I1004 05:31:06.488018 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-67msj" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="registry-server" containerID="cri-o://710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334" gracePeriod=2 Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.493436 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.498322 4992 generic.go:334] "Generic (PLEG): container finished" podID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerID="710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334" exitCode=0 Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.498398 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67msj" event={"ID":"3e8f0b2e-d98a-4939-88b5-7cb82bacb692","Type":"ContainerDied","Data":"710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334"} Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.498461 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67msj" event={"ID":"3e8f0b2e-d98a-4939-88b5-7cb82bacb692","Type":"ContainerDied","Data":"485a0c18b6e9e376d52af54280772ec15a4228be40c121abe5a740e3c2d8f75a"} Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.498487 4992 scope.go:117] "RemoveContainer" containerID="710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.498537 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67msj" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.506731 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-utilities\") pod \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.506828 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfgw5\" (UniqueName: \"kubernetes.io/projected/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-kube-api-access-wfgw5\") pod \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.506999 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-catalog-content\") pod \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\" (UID: \"3e8f0b2e-d98a-4939-88b5-7cb82bacb692\") " Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.508184 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-utilities" (OuterVolumeSpecName: "utilities") pod "3e8f0b2e-d98a-4939-88b5-7cb82bacb692" (UID: "3e8f0b2e-d98a-4939-88b5-7cb82bacb692"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.512688 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-kube-api-access-wfgw5" (OuterVolumeSpecName: "kube-api-access-wfgw5") pod "3e8f0b2e-d98a-4939-88b5-7cb82bacb692" (UID: "3e8f0b2e-d98a-4939-88b5-7cb82bacb692"). InnerVolumeSpecName "kube-api-access-wfgw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.547171 4992 scope.go:117] "RemoveContainer" containerID="e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.584135 4992 scope.go:117] "RemoveContainer" containerID="1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.609790 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.609823 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfgw5\" (UniqueName: \"kubernetes.io/projected/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-kube-api-access-wfgw5\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.626848 4992 scope.go:117] "RemoveContainer" containerID="710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334" Oct 04 05:31:07 crc kubenswrapper[4992]: E1004 05:31:07.627267 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334\": container with ID starting with 710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334 not found: ID does not exist" containerID="710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.627314 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334"} err="failed to get container status \"710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334\": rpc error: code = NotFound desc = could not find container \"710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334\": container with ID starting with 710480aba1181327525f390cdf0c9b899768dfe26e206eebca8074abf1376334 not found: ID does not exist" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.627346 4992 scope.go:117] "RemoveContainer" containerID="e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528" Oct 04 05:31:07 crc kubenswrapper[4992]: E1004 05:31:07.627646 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528\": container with ID starting with e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528 not found: ID does not exist" containerID="e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.627687 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528"} err="failed to get container status \"e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528\": rpc error: code = NotFound desc = could not find container \"e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528\": container with ID starting with e3c4f79d70fe2ec2b4d3cd20708f4b81aeb4722b5cfea295b697c246dd31e528 not found: ID does not exist" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.627709 4992 scope.go:117] "RemoveContainer" containerID="1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b" Oct 04 05:31:07 crc kubenswrapper[4992]: E1004 05:31:07.628014 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b\": container with ID starting with 1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b not found: ID does not exist" containerID="1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.628032 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b"} err="failed to get container status \"1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b\": rpc error: code = NotFound desc = could not find container \"1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b\": container with ID starting with 1d74b451f5d2fbf1dee8dff2098ee720b9c7569009bbf126987f157dc89cc82b not found: ID does not exist" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.631087 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e8f0b2e-d98a-4939-88b5-7cb82bacb692" (UID: "3e8f0b2e-d98a-4939-88b5-7cb82bacb692"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.711759 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8f0b2e-d98a-4939-88b5-7cb82bacb692-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.832591 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67msj"] Oct 04 05:31:07 crc kubenswrapper[4992]: I1004 05:31:07.842397 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-67msj"] Oct 04 05:31:08 crc kubenswrapper[4992]: I1004 05:31:08.330311 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" path="/var/lib/kubelet/pods/3e8f0b2e-d98a-4939-88b5-7cb82bacb692/volumes" Oct 04 05:31:13 crc kubenswrapper[4992]: I1004 05:31:13.044982 4992 scope.go:117] "RemoveContainer" containerID="89d646e67c891a5fb50ea641c0ae6168148cf2c34cc8cbc665bd62ce609f59c9" Oct 04 05:31:13 crc kubenswrapper[4992]: I1004 05:31:13.070050 4992 scope.go:117] "RemoveContainer" containerID="fee0f62e350126841a91f76687a950b6dc2338a7c839d67cbfca66afa568f2ae" Oct 04 05:31:13 crc kubenswrapper[4992]: I1004 05:31:13.102975 4992 scope.go:117] "RemoveContainer" containerID="7a586454bc1eba450665cc764f4f77352a045cdedc5a8b9d0b56936893033642" Oct 04 05:32:50 crc kubenswrapper[4992]: I1004 05:32:50.894593 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:32:50 crc kubenswrapper[4992]: I1004 05:32:50.895199 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:33:20 crc kubenswrapper[4992]: I1004 05:33:20.894662 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:33:20 crc kubenswrapper[4992]: I1004 05:33:20.895034 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:33:50 crc kubenswrapper[4992]: I1004 05:33:50.894967 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:33:50 crc kubenswrapper[4992]: I1004 05:33:50.895781 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:33:50 crc kubenswrapper[4992]: I1004 05:33:50.895856 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 05:33:50 crc kubenswrapper[4992]: I1004 05:33:50.897021 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95a2327fa1567e14683aac6a2334df02751530ba6b438478d3fecaa93c2aa785"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:33:50 crc kubenswrapper[4992]: I1004 05:33:50.897130 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://95a2327fa1567e14683aac6a2334df02751530ba6b438478d3fecaa93c2aa785" gracePeriod=600 Oct 04 05:33:56 crc kubenswrapper[4992]: I1004 05:33:56.842750 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="6d24bc2b-9416-4bf0-aff1-081826b9c9d9" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.166:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:34:01 crc kubenswrapper[4992]: I1004 05:34:01.722815 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 04 05:34:01 crc kubenswrapper[4992]: I1004 05:34:01.885661 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="6d24bc2b-9416-4bf0-aff1-081826b9c9d9" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.166:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:34:06 crc kubenswrapper[4992]: I1004 05:34:06.724815 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 04 05:34:06 crc kubenswrapper[4992]: I1004 05:34:06.926611 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="6d24bc2b-9416-4bf0-aff1-081826b9c9d9" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.166:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 05:34:06 crc kubenswrapper[4992]: I1004 05:34:06.926708 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 05:34:06 crc kubenswrapper[4992]: I1004 05:34:06.927608 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"588cd67a06c1ba3fbc54d80c79d7447acffd47584775e385023f03a07bbcc77c"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed liveness probe, will be restarted" Oct 04 05:34:06 crc kubenswrapper[4992]: I1004 05:34:06.927681 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6d24bc2b-9416-4bf0-aff1-081826b9c9d9" containerName="cinder-scheduler" containerID="cri-o://588cd67a06c1ba3fbc54d80c79d7447acffd47584775e385023f03a07bbcc77c" gracePeriod=30 Oct 04 05:34:11 crc kubenswrapper[4992]: I1004 05:34:11.729191 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Oct 04 05:34:11 crc kubenswrapper[4992]: I1004 05:34:11.729870 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Oct 04 05:34:11 crc kubenswrapper[4992]: I1004 05:34:11.730648 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"715c12561023a725df6254da792b9cc8bdcf7f098a7935409b169e6d19f5cf5b"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Oct 04 05:34:11 crc kubenswrapper[4992]: I1004 05:34:11.730768 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerName="ceilometer-central-agent" containerID="cri-o://715c12561023a725df6254da792b9cc8bdcf7f098a7935409b169e6d19f5cf5b" gracePeriod=30 Oct 04 05:34:26 crc kubenswrapper[4992]: I1004 05:34:26.721093 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Oct 04 05:34:56 crc kubenswrapper[4992]: I1004 05:34:56.723152 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.992081 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ftf65"] Oct 04 05:35:01 crc kubenswrapper[4992]: E1004 05:35:01.993486 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="extract-utilities" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.993518 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="extract-utilities" Oct 04 05:35:01 crc kubenswrapper[4992]: E1004 05:35:01.993578 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e208c780-5970-45e0-adce-55980aa4bd88" containerName="registry-server" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.993596 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="e208c780-5970-45e0-adce-55980aa4bd88" containerName="registry-server" Oct 04 05:35:01 crc kubenswrapper[4992]: E1004 05:35:01.993627 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="extract-content" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.993640 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="extract-content" Oct 04 05:35:01 crc kubenswrapper[4992]: E1004 05:35:01.993658 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e208c780-5970-45e0-adce-55980aa4bd88" containerName="extract-content" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.993670 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="e208c780-5970-45e0-adce-55980aa4bd88" containerName="extract-content" Oct 04 05:35:01 crc kubenswrapper[4992]: E1004 05:35:01.993693 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e208c780-5970-45e0-adce-55980aa4bd88" containerName="extract-utilities" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.993705 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="e208c780-5970-45e0-adce-55980aa4bd88" containerName="extract-utilities" Oct 04 05:35:01 crc kubenswrapper[4992]: E1004 05:35:01.993736 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="registry-server" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.993748 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="registry-server" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.994081 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8f0b2e-d98a-4939-88b5-7cb82bacb692" containerName="registry-server" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.994106 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="e208c780-5970-45e0-adce-55980aa4bd88" containerName="registry-server" Oct 04 05:35:01 crc kubenswrapper[4992]: I1004 05:35:01.996806 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.038148 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftf65"] Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.106588 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-catalog-content\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.106697 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-utilities\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.106723 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njm76\" (UniqueName: \"kubernetes.io/projected/86561d96-9a35-4b25-b13b-703cae5775a4-kube-api-access-njm76\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.208950 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-utilities\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.209014 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njm76\" (UniqueName: \"kubernetes.io/projected/86561d96-9a35-4b25-b13b-703cae5775a4-kube-api-access-njm76\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.209146 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-catalog-content\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.209736 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-utilities\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.209754 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-catalog-content\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.227673 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njm76\" (UniqueName: \"kubernetes.io/projected/86561d96-9a35-4b25-b13b-703cae5775a4-kube-api-access-njm76\") pod \"redhat-marketplace-ftf65\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.344672 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:02 crc kubenswrapper[4992]: I1004 05:35:02.850326 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftf65"] Oct 04 05:35:09 crc kubenswrapper[4992]: I1004 05:35:09.427093 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_machine-config-daemon-p5l9q_d6d4bf54-7f00-4c33-830c-8368dd9678df/machine-config-daemon/12.log" Oct 04 05:35:09 crc kubenswrapper[4992]: I1004 05:35:09.429118 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="95a2327fa1567e14683aac6a2334df02751530ba6b438478d3fecaa93c2aa785" exitCode=-1 Oct 04 05:35:09 crc kubenswrapper[4992]: I1004 05:35:09.430150 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"95a2327fa1567e14683aac6a2334df02751530ba6b438478d3fecaa93c2aa785"} Oct 04 05:35:09 crc kubenswrapper[4992]: I1004 05:35:09.430193 4992 scope.go:117] "RemoveContainer" containerID="50313b15d20f039d4dee1831c150e173a7f8009b061b404bb95cdb47097fc514" Oct 04 05:35:10 crc kubenswrapper[4992]: I1004 05:35:10.438391 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftf65" event={"ID":"86561d96-9a35-4b25-b13b-703cae5775a4","Type":"ContainerStarted","Data":"0b51b282532616346ea1c2d9c2cc037b1c3649e92bd688207dda6e7b2cf671d0"} Oct 04 05:35:10 crc kubenswrapper[4992]: I1004 05:35:10.438693 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftf65" event={"ID":"86561d96-9a35-4b25-b13b-703cae5775a4","Type":"ContainerStarted","Data":"0096f9376ff8a0bf173df7f34aca95693620035b29ccd819c4e10b0696025d2c"} Oct 04 05:35:10 crc kubenswrapper[4992]: I1004 05:35:10.440586 4992 generic.go:334] "Generic (PLEG): container finished" podID="6d24bc2b-9416-4bf0-aff1-081826b9c9d9" containerID="588cd67a06c1ba3fbc54d80c79d7447acffd47584775e385023f03a07bbcc77c" exitCode=137 Oct 04 05:35:10 crc kubenswrapper[4992]: I1004 05:35:10.441432 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6d24bc2b-9416-4bf0-aff1-081826b9c9d9","Type":"ContainerDied","Data":"588cd67a06c1ba3fbc54d80c79d7447acffd47584775e385023f03a07bbcc77c"} Oct 04 05:35:11 crc kubenswrapper[4992]: I1004 05:35:11.454297 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8"} Oct 04 05:35:11 crc kubenswrapper[4992]: I1004 05:35:11.457881 4992 generic.go:334] "Generic (PLEG): container finished" podID="86561d96-9a35-4b25-b13b-703cae5775a4" containerID="0b51b282532616346ea1c2d9c2cc037b1c3649e92bd688207dda6e7b2cf671d0" exitCode=0 Oct 04 05:35:11 crc kubenswrapper[4992]: I1004 05:35:11.457924 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftf65" event={"ID":"86561d96-9a35-4b25-b13b-703cae5775a4","Type":"ContainerDied","Data":"0b51b282532616346ea1c2d9c2cc037b1c3649e92bd688207dda6e7b2cf671d0"} Oct 04 05:35:12 crc kubenswrapper[4992]: I1004 05:35:12.469792 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6d24bc2b-9416-4bf0-aff1-081826b9c9d9","Type":"ContainerStarted","Data":"5b818e5e823ec6c89c5bca28e86a5fb569cf7bd8be31abe5454570f7da1f27e4"} Oct 04 05:35:13 crc kubenswrapper[4992]: I1004 05:35:13.478563 4992 generic.go:334] "Generic (PLEG): container finished" podID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerID="715c12561023a725df6254da792b9cc8bdcf7f098a7935409b169e6d19f5cf5b" exitCode=0 Oct 04 05:35:13 crc kubenswrapper[4992]: I1004 05:35:13.478629 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerDied","Data":"715c12561023a725df6254da792b9cc8bdcf7f098a7935409b169e6d19f5cf5b"} Oct 04 05:35:14 crc kubenswrapper[4992]: I1004 05:35:14.801405 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 05:35:15 crc kubenswrapper[4992]: I1004 05:35:15.499194 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerStarted","Data":"ceb7fb069baec0264a9b90e60a7a4d35ec6cc176822c535a95952d8c03e21df4"} Oct 04 05:35:19 crc kubenswrapper[4992]: I1004 05:35:19.536779 4992 generic.go:334] "Generic (PLEG): container finished" podID="86561d96-9a35-4b25-b13b-703cae5775a4" containerID="858c7c6f976cc95a9e1692173766867bd9df84fd7139a40950b7d94ce08494e8" exitCode=0 Oct 04 05:35:19 crc kubenswrapper[4992]: I1004 05:35:19.536876 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftf65" event={"ID":"86561d96-9a35-4b25-b13b-703cae5775a4","Type":"ContainerDied","Data":"858c7c6f976cc95a9e1692173766867bd9df84fd7139a40950b7d94ce08494e8"} Oct 04 05:35:19 crc kubenswrapper[4992]: I1004 05:35:19.813047 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 05:35:21 crc kubenswrapper[4992]: I1004 05:35:21.557879 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftf65" event={"ID":"86561d96-9a35-4b25-b13b-703cae5775a4","Type":"ContainerStarted","Data":"878d1e764a89d6737e28b0ce8e8d10fa69246394bb16cde6d4846c79550bd48e"} Oct 04 05:35:21 crc kubenswrapper[4992]: I1004 05:35:21.577275 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ftf65" podStartSLOduration=11.532918113000001 podStartE2EDuration="20.57725523s" podCreationTimestamp="2025-10-04 05:35:01 +0000 UTC" firstStartedPulling="2025-10-04 05:35:11.459399625 +0000 UTC m=+3745.307075093" lastFinishedPulling="2025-10-04 05:35:20.503736742 +0000 UTC m=+3754.351412210" observedRunningTime="2025-10-04 05:35:21.575189224 +0000 UTC m=+3755.422864702" watchObservedRunningTime="2025-10-04 05:35:21.57725523 +0000 UTC m=+3755.424930698" Oct 04 05:35:21 crc kubenswrapper[4992]: I1004 05:35:21.772691 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerName="ceilometer-notification-agent" probeResult="failure" output=< Oct 04 05:35:21 crc kubenswrapper[4992]: Unkown error: Expecting value: line 1 column 1 (char 0) Oct 04 05:35:21 crc kubenswrapper[4992]: > Oct 04 05:35:21 crc kubenswrapper[4992]: I1004 05:35:21.772781 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Oct 04 05:35:21 crc kubenswrapper[4992]: I1004 05:35:21.773600 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-notification-agent" containerStatusID={"Type":"cri-o","ID":"99cafd3f4d325875ed8a631e1d05e044cf4ecd9bbab472ce70775ba5ee5c1aea"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-notification-agent failed liveness probe, will be restarted" Oct 04 05:35:21 crc kubenswrapper[4992]: I1004 05:35:21.773668 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerName="ceilometer-notification-agent" containerID="cri-o://99cafd3f4d325875ed8a631e1d05e044cf4ecd9bbab472ce70775ba5ee5c1aea" gracePeriod=30 Oct 04 05:35:22 crc kubenswrapper[4992]: I1004 05:35:22.345039 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:22 crc kubenswrapper[4992]: I1004 05:35:22.345085 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:22 crc kubenswrapper[4992]: I1004 05:35:22.395111 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:23 crc kubenswrapper[4992]: I1004 05:35:23.579604 4992 generic.go:334] "Generic (PLEG): container finished" podID="bc693b9c-b61e-4794-b736-59f588ee80a1" containerID="99cafd3f4d325875ed8a631e1d05e044cf4ecd9bbab472ce70775ba5ee5c1aea" exitCode=0 Oct 04 05:35:23 crc kubenswrapper[4992]: I1004 05:35:23.579664 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerDied","Data":"99cafd3f4d325875ed8a631e1d05e044cf4ecd9bbab472ce70775ba5ee5c1aea"} Oct 04 05:35:24 crc kubenswrapper[4992]: I1004 05:35:24.591617 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc693b9c-b61e-4794-b736-59f588ee80a1","Type":"ContainerStarted","Data":"5ac692cc8f34e6d163adf2965911961f951cfbcbd263815f6a2f7fbaa3ace4df"} Oct 04 05:35:32 crc kubenswrapper[4992]: I1004 05:35:32.392690 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:32 crc kubenswrapper[4992]: I1004 05:35:32.449072 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftf65"] Oct 04 05:35:32 crc kubenswrapper[4992]: I1004 05:35:32.665832 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ftf65" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" containerName="registry-server" containerID="cri-o://878d1e764a89d6737e28b0ce8e8d10fa69246394bb16cde6d4846c79550bd48e" gracePeriod=2 Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.693619 4992 generic.go:334] "Generic (PLEG): container finished" podID="86561d96-9a35-4b25-b13b-703cae5775a4" containerID="878d1e764a89d6737e28b0ce8e8d10fa69246394bb16cde6d4846c79550bd48e" exitCode=0 Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.693673 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftf65" event={"ID":"86561d96-9a35-4b25-b13b-703cae5775a4","Type":"ContainerDied","Data":"878d1e764a89d6737e28b0ce8e8d10fa69246394bb16cde6d4846c79550bd48e"} Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.883931 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.945420 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-catalog-content\") pod \"86561d96-9a35-4b25-b13b-703cae5775a4\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.945535 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-utilities\") pod \"86561d96-9a35-4b25-b13b-703cae5775a4\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.945692 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njm76\" (UniqueName: \"kubernetes.io/projected/86561d96-9a35-4b25-b13b-703cae5775a4-kube-api-access-njm76\") pod \"86561d96-9a35-4b25-b13b-703cae5775a4\" (UID: \"86561d96-9a35-4b25-b13b-703cae5775a4\") " Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.946401 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-utilities" (OuterVolumeSpecName: "utilities") pod "86561d96-9a35-4b25-b13b-703cae5775a4" (UID: "86561d96-9a35-4b25-b13b-703cae5775a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.951459 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86561d96-9a35-4b25-b13b-703cae5775a4-kube-api-access-njm76" (OuterVolumeSpecName: "kube-api-access-njm76") pod "86561d96-9a35-4b25-b13b-703cae5775a4" (UID: "86561d96-9a35-4b25-b13b-703cae5775a4"). InnerVolumeSpecName "kube-api-access-njm76". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:35:33 crc kubenswrapper[4992]: I1004 05:35:33.958276 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86561d96-9a35-4b25-b13b-703cae5775a4" (UID: "86561d96-9a35-4b25-b13b-703cae5775a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.048081 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.048140 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njm76\" (UniqueName: \"kubernetes.io/projected/86561d96-9a35-4b25-b13b-703cae5775a4-kube-api-access-njm76\") on node \"crc\" DevicePath \"\"" Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.048163 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86561d96-9a35-4b25-b13b-703cae5775a4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.705865 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftf65" event={"ID":"86561d96-9a35-4b25-b13b-703cae5775a4","Type":"ContainerDied","Data":"0096f9376ff8a0bf173df7f34aca95693620035b29ccd819c4e10b0696025d2c"} Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.705904 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftf65" Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.705925 4992 scope.go:117] "RemoveContainer" containerID="878d1e764a89d6737e28b0ce8e8d10fa69246394bb16cde6d4846c79550bd48e" Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.731025 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftf65"] Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.731767 4992 scope.go:117] "RemoveContainer" containerID="858c7c6f976cc95a9e1692173766867bd9df84fd7139a40950b7d94ce08494e8" Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.741845 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftf65"] Oct 04 05:35:34 crc kubenswrapper[4992]: I1004 05:35:34.753077 4992 scope.go:117] "RemoveContainer" containerID="0b51b282532616346ea1c2d9c2cc037b1c3649e92bd688207dda6e7b2cf671d0" Oct 04 05:35:36 crc kubenswrapper[4992]: I1004 05:35:36.350573 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" path="/var/lib/kubelet/pods/86561d96-9a35-4b25-b13b-703cae5775a4/volumes" Oct 04 05:36:56 crc kubenswrapper[4992]: I1004 05:36:56.505253 4992 generic.go:334] "Generic (PLEG): container finished" podID="dab26aab-5877-4a73-9433-7f76c31377dc" containerID="e4780c955e26f441d9867a104d5e2368386bdbcd4837f20611f06b2d576a0964" exitCode=1 Oct 04 05:36:56 crc kubenswrapper[4992]: I1004 05:36:56.505412 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"dab26aab-5877-4a73-9433-7f76c31377dc","Type":"ContainerDied","Data":"e4780c955e26f441d9867a104d5e2368386bdbcd4837f20611f06b2d576a0964"} Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.897967 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.928906 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929000 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-config-data\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929042 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-workdir\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929073 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-temporary\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929131 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ca-certs\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929174 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929706 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929752 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config-secret\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929789 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt9bw\" (UniqueName: \"kubernetes.io/projected/dab26aab-5877-4a73-9433-7f76c31377dc-kube-api-access-mt9bw\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929859 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ssh-key\") pod \"dab26aab-5877-4a73-9433-7f76c31377dc\" (UID: \"dab26aab-5877-4a73-9433-7f76c31377dc\") " Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.930225 4992 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.929853 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-config-data" (OuterVolumeSpecName: "config-data") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.932268 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.935785 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dab26aab-5877-4a73-9433-7f76c31377dc-kube-api-access-mt9bw" (OuterVolumeSpecName: "kube-api-access-mt9bw") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "kube-api-access-mt9bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.935869 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.954767 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.957860 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.962119 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:36:57 crc kubenswrapper[4992]: I1004 05:36:57.980446 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "dab26aab-5877-4a73-9433-7f76c31377dc" (UID: "dab26aab-5877-4a73-9433-7f76c31377dc"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.032174 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt9bw\" (UniqueName: \"kubernetes.io/projected/dab26aab-5877-4a73-9433-7f76c31377dc-kube-api-access-mt9bw\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.032217 4992 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.032234 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.032245 4992 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dab26aab-5877-4a73-9433-7f76c31377dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.032258 4992 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/dab26aab-5877-4a73-9433-7f76c31377dc-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.032271 4992 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.032312 4992 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.032324 4992 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dab26aab-5877-4a73-9433-7f76c31377dc-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.068623 4992 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.133950 4992 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.523515 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"dab26aab-5877-4a73-9433-7f76c31377dc","Type":"ContainerDied","Data":"f3491616396043d01cf8d840d8e108a0f756c58357a2dc20456013681c6b93aa"} Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.523564 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3491616396043d01cf8d840d8e108a0f756c58357a2dc20456013681c6b93aa" Oct 04 05:36:58 crc kubenswrapper[4992]: I1004 05:36:58.523636 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.258411 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 05:37:02 crc kubenswrapper[4992]: E1004 05:37:02.259439 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dab26aab-5877-4a73-9433-7f76c31377dc" containerName="tempest-tests-tempest-tests-runner" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.259456 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="dab26aab-5877-4a73-9433-7f76c31377dc" containerName="tempest-tests-tempest-tests-runner" Oct 04 05:37:02 crc kubenswrapper[4992]: E1004 05:37:02.259484 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" containerName="extract-utilities" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.259492 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" containerName="extract-utilities" Oct 04 05:37:02 crc kubenswrapper[4992]: E1004 05:37:02.259508 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" containerName="extract-content" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.259516 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" containerName="extract-content" Oct 04 05:37:02 crc kubenswrapper[4992]: E1004 05:37:02.259533 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" containerName="registry-server" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.259541 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" containerName="registry-server" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.259773 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="dab26aab-5877-4a73-9433-7f76c31377dc" containerName="tempest-tests-tempest-tests-runner" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.259790 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="86561d96-9a35-4b25-b13b-703cae5775a4" containerName="registry-server" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.260604 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.262950 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pr8fk" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.269786 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.311703 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nxhj\" (UniqueName: \"kubernetes.io/projected/eb71ad30-b074-45c1-a11a-42eba81438f6-kube-api-access-2nxhj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb71ad30-b074-45c1-a11a-42eba81438f6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.312469 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb71ad30-b074-45c1-a11a-42eba81438f6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.414645 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb71ad30-b074-45c1-a11a-42eba81438f6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.415165 4992 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb71ad30-b074-45c1-a11a-42eba81438f6\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.415336 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nxhj\" (UniqueName: \"kubernetes.io/projected/eb71ad30-b074-45c1-a11a-42eba81438f6-kube-api-access-2nxhj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb71ad30-b074-45c1-a11a-42eba81438f6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.437352 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nxhj\" (UniqueName: \"kubernetes.io/projected/eb71ad30-b074-45c1-a11a-42eba81438f6-kube-api-access-2nxhj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb71ad30-b074-45c1-a11a-42eba81438f6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.442812 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb71ad30-b074-45c1-a11a-42eba81438f6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:02 crc kubenswrapper[4992]: I1004 05:37:02.596248 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 05:37:03 crc kubenswrapper[4992]: I1004 05:37:03.062224 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 05:37:03 crc kubenswrapper[4992]: I1004 05:37:03.064291 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:37:03 crc kubenswrapper[4992]: I1004 05:37:03.572842 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"eb71ad30-b074-45c1-a11a-42eba81438f6","Type":"ContainerStarted","Data":"5402fac28f54db26a99624f413e9bfb0bc5b4b0eb8ca923f0ce799fc2f61c687"} Oct 04 05:37:05 crc kubenswrapper[4992]: I1004 05:37:05.598281 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"eb71ad30-b074-45c1-a11a-42eba81438f6","Type":"ContainerStarted","Data":"9a1b6b4cca05695c6c5311f339eaf785bfa6f612e2f0a42b16a3b54675ac7dc0"} Oct 04 05:37:05 crc kubenswrapper[4992]: I1004 05:37:05.619385 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.751319761 podStartE2EDuration="3.619348449s" podCreationTimestamp="2025-10-04 05:37:02 +0000 UTC" firstStartedPulling="2025-10-04 05:37:03.063984486 +0000 UTC m=+3856.911659964" lastFinishedPulling="2025-10-04 05:37:04.932013174 +0000 UTC m=+3858.779688652" observedRunningTime="2025-10-04 05:37:05.610241155 +0000 UTC m=+3859.457916633" watchObservedRunningTime="2025-10-04 05:37:05.619348449 +0000 UTC m=+3859.467023917" Oct 04 05:37:20 crc kubenswrapper[4992]: I1004 05:37:20.894713 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:37:20 crc kubenswrapper[4992]: I1004 05:37:20.895312 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:37:27 crc kubenswrapper[4992]: I1004 05:37:27.909859 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-df5cv/must-gather-8v9r4"] Oct 04 05:37:27 crc kubenswrapper[4992]: I1004 05:37:27.911852 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:37:27 crc kubenswrapper[4992]: I1004 05:37:27.914454 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-df5cv"/"openshift-service-ca.crt" Oct 04 05:37:27 crc kubenswrapper[4992]: I1004 05:37:27.914565 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-df5cv"/"default-dockercfg-66nwm" Oct 04 05:37:27 crc kubenswrapper[4992]: I1004 05:37:27.916873 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-df5cv"/"kube-root-ca.crt" Oct 04 05:37:27 crc kubenswrapper[4992]: I1004 05:37:27.922520 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-df5cv/must-gather-8v9r4"] Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.041411 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lth5\" (UniqueName: \"kubernetes.io/projected/4cc5e9e6-5741-48ee-97aa-c694705babf5-kube-api-access-9lth5\") pod \"must-gather-8v9r4\" (UID: \"4cc5e9e6-5741-48ee-97aa-c694705babf5\") " pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.041480 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4cc5e9e6-5741-48ee-97aa-c694705babf5-must-gather-output\") pod \"must-gather-8v9r4\" (UID: \"4cc5e9e6-5741-48ee-97aa-c694705babf5\") " pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.142816 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lth5\" (UniqueName: \"kubernetes.io/projected/4cc5e9e6-5741-48ee-97aa-c694705babf5-kube-api-access-9lth5\") pod \"must-gather-8v9r4\" (UID: \"4cc5e9e6-5741-48ee-97aa-c694705babf5\") " pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.142890 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4cc5e9e6-5741-48ee-97aa-c694705babf5-must-gather-output\") pod \"must-gather-8v9r4\" (UID: \"4cc5e9e6-5741-48ee-97aa-c694705babf5\") " pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.143490 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4cc5e9e6-5741-48ee-97aa-c694705babf5-must-gather-output\") pod \"must-gather-8v9r4\" (UID: \"4cc5e9e6-5741-48ee-97aa-c694705babf5\") " pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.162499 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lth5\" (UniqueName: \"kubernetes.io/projected/4cc5e9e6-5741-48ee-97aa-c694705babf5-kube-api-access-9lth5\") pod \"must-gather-8v9r4\" (UID: \"4cc5e9e6-5741-48ee-97aa-c694705babf5\") " pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.234615 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.701490 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-df5cv/must-gather-8v9r4"] Oct 04 05:37:28 crc kubenswrapper[4992]: W1004 05:37:28.705760 4992 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cc5e9e6_5741_48ee_97aa_c694705babf5.slice/crio-2d12a84c0fda2ab1b67a06b2ab556583a61a95590d23840d93a073c5996c5a38 WatchSource:0}: Error finding container 2d12a84c0fda2ab1b67a06b2ab556583a61a95590d23840d93a073c5996c5a38: Status 404 returned error can't find the container with id 2d12a84c0fda2ab1b67a06b2ab556583a61a95590d23840d93a073c5996c5a38 Oct 04 05:37:28 crc kubenswrapper[4992]: I1004 05:37:28.827422 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/must-gather-8v9r4" event={"ID":"4cc5e9e6-5741-48ee-97aa-c694705babf5","Type":"ContainerStarted","Data":"2d12a84c0fda2ab1b67a06b2ab556583a61a95590d23840d93a073c5996c5a38"} Oct 04 05:37:35 crc kubenswrapper[4992]: I1004 05:37:35.916280 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/must-gather-8v9r4" event={"ID":"4cc5e9e6-5741-48ee-97aa-c694705babf5","Type":"ContainerStarted","Data":"2481c6e385e3c1770500933799ebfe1a0ae3b3478ee2dbc38ca2d149445a3bb9"} Oct 04 05:37:35 crc kubenswrapper[4992]: I1004 05:37:35.916932 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/must-gather-8v9r4" event={"ID":"4cc5e9e6-5741-48ee-97aa-c694705babf5","Type":"ContainerStarted","Data":"6a3292ee2b54647420b2fc86a26f0c289fe04cd9de8302eb2a2a5740015baa67"} Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.621555 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-df5cv/must-gather-8v9r4" podStartSLOduration=5.707194159 podStartE2EDuration="11.621536809s" podCreationTimestamp="2025-10-04 05:37:27 +0000 UTC" firstStartedPulling="2025-10-04 05:37:28.707615507 +0000 UTC m=+3882.555290975" lastFinishedPulling="2025-10-04 05:37:34.621958157 +0000 UTC m=+3888.469633625" observedRunningTime="2025-10-04 05:37:35.936395276 +0000 UTC m=+3889.784070744" watchObservedRunningTime="2025-10-04 05:37:38.621536809 +0000 UTC m=+3892.469212277" Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.623495 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-df5cv/crc-debug-wgkrl"] Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.624909 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.762528 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74kz6\" (UniqueName: \"kubernetes.io/projected/5f7b9547-55e2-4112-be8d-28191704a7c2-kube-api-access-74kz6\") pod \"crc-debug-wgkrl\" (UID: \"5f7b9547-55e2-4112-be8d-28191704a7c2\") " pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.762750 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f7b9547-55e2-4112-be8d-28191704a7c2-host\") pod \"crc-debug-wgkrl\" (UID: \"5f7b9547-55e2-4112-be8d-28191704a7c2\") " pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.864441 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74kz6\" (UniqueName: \"kubernetes.io/projected/5f7b9547-55e2-4112-be8d-28191704a7c2-kube-api-access-74kz6\") pod \"crc-debug-wgkrl\" (UID: \"5f7b9547-55e2-4112-be8d-28191704a7c2\") " pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.864736 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f7b9547-55e2-4112-be8d-28191704a7c2-host\") pod \"crc-debug-wgkrl\" (UID: \"5f7b9547-55e2-4112-be8d-28191704a7c2\") " pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.866600 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f7b9547-55e2-4112-be8d-28191704a7c2-host\") pod \"crc-debug-wgkrl\" (UID: \"5f7b9547-55e2-4112-be8d-28191704a7c2\") " pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.885816 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74kz6\" (UniqueName: \"kubernetes.io/projected/5f7b9547-55e2-4112-be8d-28191704a7c2-kube-api-access-74kz6\") pod \"crc-debug-wgkrl\" (UID: \"5f7b9547-55e2-4112-be8d-28191704a7c2\") " pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:37:38 crc kubenswrapper[4992]: I1004 05:37:38.944890 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:37:39 crc kubenswrapper[4992]: I1004 05:37:39.976998 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/crc-debug-wgkrl" event={"ID":"5f7b9547-55e2-4112-be8d-28191704a7c2","Type":"ContainerStarted","Data":"dee7214a6b594f087941e93863c7c25e6416c19d96b33575e3a571937604a855"} Oct 04 05:37:50 crc kubenswrapper[4992]: I1004 05:37:50.894871 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:37:50 crc kubenswrapper[4992]: I1004 05:37:50.895414 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:38:00 crc kubenswrapper[4992]: I1004 05:38:00.218656 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/crc-debug-wgkrl" event={"ID":"5f7b9547-55e2-4112-be8d-28191704a7c2","Type":"ContainerStarted","Data":"28ae4c509a0f46149029a591f04c1040344b5f1c58f1331dd8d0b02960c55680"} Oct 04 05:38:00 crc kubenswrapper[4992]: I1004 05:38:00.240957 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-df5cv/crc-debug-wgkrl" podStartSLOduration=2.240749176 podStartE2EDuration="22.240916589s" podCreationTimestamp="2025-10-04 05:37:38 +0000 UTC" firstStartedPulling="2025-10-04 05:37:39.022130458 +0000 UTC m=+3892.869805926" lastFinishedPulling="2025-10-04 05:37:59.022297871 +0000 UTC m=+3912.869973339" observedRunningTime="2025-10-04 05:38:00.232668069 +0000 UTC m=+3914.080343537" watchObservedRunningTime="2025-10-04 05:38:00.240916589 +0000 UTC m=+3914.088592047" Oct 04 05:38:20 crc kubenswrapper[4992]: I1004 05:38:20.895033 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:38:20 crc kubenswrapper[4992]: I1004 05:38:20.895799 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 05:38:20 crc kubenswrapper[4992]: I1004 05:38:20.895877 4992 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" Oct 04 05:38:20 crc kubenswrapper[4992]: I1004 05:38:20.897022 4992 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8"} pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 05:38:20 crc kubenswrapper[4992]: I1004 05:38:20.897094 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" containerID="cri-o://16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" gracePeriod=600 Oct 04 05:38:21 crc kubenswrapper[4992]: E1004 05:38:21.030970 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:38:21 crc kubenswrapper[4992]: I1004 05:38:21.406897 4992 generic.go:334] "Generic (PLEG): container finished" podID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" exitCode=0 Oct 04 05:38:21 crc kubenswrapper[4992]: I1004 05:38:21.406948 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerDied","Data":"16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8"} Oct 04 05:38:21 crc kubenswrapper[4992]: I1004 05:38:21.407009 4992 scope.go:117] "RemoveContainer" containerID="95a2327fa1567e14683aac6a2334df02751530ba6b438478d3fecaa93c2aa785" Oct 04 05:38:21 crc kubenswrapper[4992]: I1004 05:38:21.408110 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:38:21 crc kubenswrapper[4992]: E1004 05:38:21.408434 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:38:37 crc kubenswrapper[4992]: I1004 05:38:37.320053 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:38:37 crc kubenswrapper[4992]: E1004 05:38:37.321456 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:38:41 crc kubenswrapper[4992]: I1004 05:38:41.339207 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5dfd9d5cfd-4mzm5_acb287d0-f746-4d74-acaa-4153228f3194/barbican-api-log/0.log" Oct 04 05:38:41 crc kubenswrapper[4992]: I1004 05:38:41.355339 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5dfd9d5cfd-4mzm5_acb287d0-f746-4d74-acaa-4153228f3194/barbican-api/0.log" Oct 04 05:38:41 crc kubenswrapper[4992]: I1004 05:38:41.534203 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5ff4b555c4-vn5np_a4b0448a-9b84-41c8-a87f-fbc17f92411b/barbican-keystone-listener/0.log" Oct 04 05:38:41 crc kubenswrapper[4992]: I1004 05:38:41.637680 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5ff4b555c4-vn5np_a4b0448a-9b84-41c8-a87f-fbc17f92411b/barbican-keystone-listener-log/0.log" Oct 04 05:38:41 crc kubenswrapper[4992]: I1004 05:38:41.747267 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-775b66879-7sgzw_70af60aa-3bee-44bf-874e-648aaefe032d/barbican-worker/0.log" Oct 04 05:38:41 crc kubenswrapper[4992]: I1004 05:38:41.845204 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-775b66879-7sgzw_70af60aa-3bee-44bf-874e-648aaefe032d/barbican-worker-log/0.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.022982 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-wgczg_645bddf3-cf63-40fc-b847-40c720339f80/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.251199 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bc693b9c-b61e-4794-b736-59f588ee80a1/ceilometer-central-agent/1.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.342798 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bc693b9c-b61e-4794-b736-59f588ee80a1/ceilometer-central-agent/0.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.419673 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bc693b9c-b61e-4794-b736-59f588ee80a1/ceilometer-notification-agent/1.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.468643 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bc693b9c-b61e-4794-b736-59f588ee80a1/ceilometer-notification-agent/0.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.550803 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bc693b9c-b61e-4794-b736-59f588ee80a1/proxy-httpd/0.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.604448 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bc693b9c-b61e-4794-b736-59f588ee80a1/sg-core/0.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.843393 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_89482934-b40f-4685-948a-a2145549ba9e/cinder-api/0.log" Oct 04 05:38:42 crc kubenswrapper[4992]: I1004 05:38:42.874634 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_89482934-b40f-4685-948a-a2145549ba9e/cinder-api-log/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.057734 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6d24bc2b-9416-4bf0-aff1-081826b9c9d9/cinder-scheduler/1.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.107637 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6d24bc2b-9416-4bf0-aff1-081826b9c9d9/cinder-scheduler/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.173596 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6d24bc2b-9416-4bf0-aff1-081826b9c9d9/probe/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.334015 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-n554v_d5fb9b36-df2c-4e2f-816c-a455cd91d3fe/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.387960 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-ptt4f_72135fbe-8e57-4d94-98f5-c9677f852e76/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.577709 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-xb8sm_5d87911a-e1f0-4c6c-a034-0b9e87faef4a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.636415 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-p2jdn_74bd057e-61c1-4538-be59-e36028dca0d9/init/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.813002 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-p2jdn_74bd057e-61c1-4538-be59-e36028dca0d9/dnsmasq-dns/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.821335 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-p2jdn_74bd057e-61c1-4538-be59-e36028dca0d9/init/0.log" Oct 04 05:38:43 crc kubenswrapper[4992]: I1004 05:38:43.878085 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-cr45c_886d41dd-15e8-48ce-bbe9-f8df87efabba/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.043946 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2b879868-2ade-447f-abc0-c6656fb65ffe/glance-httpd/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.064591 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2b879868-2ade-447f-abc0-c6656fb65ffe/glance-log/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.253212 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_28ee8ac4-5271-4bbb-85db-2a83bf7cc04a/glance-httpd/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.259795 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_28ee8ac4-5271-4bbb-85db-2a83bf7cc04a/glance-log/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.398841 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-55b5954cd6-68rnc_d75279dd-898a-4ffe-85e6-804f066de711/horizon/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.536875 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-nxgk9_3bafe4fd-0efb-4b1b-ba06-f5510110eb45/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.739957 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-55b5954cd6-68rnc_d75279dd-898a-4ffe-85e6-804f066de711/horizon-log/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.798600 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-xlrz5_9af8d0eb-c228-489d-a923-ea44104df2dd/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:44 crc kubenswrapper[4992]: I1004 05:38:44.980731 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29325901-dvngf_2fbb44fc-8c5f-4023-92dd-bbe9e7d7ce3f/keystone-cron/0.log" Oct 04 05:38:45 crc kubenswrapper[4992]: I1004 05:38:45.147119 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5bb778b4c-snbm9_e15d6dcc-af83-40dd-b5a2-936c8e5154b3/keystone-api/0.log" Oct 04 05:38:45 crc kubenswrapper[4992]: I1004 05:38:45.208264 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f056bac1-081a-4675-865b-67f6a6035b47/kube-state-metrics/0.log" Oct 04 05:38:45 crc kubenswrapper[4992]: I1004 05:38:45.319901 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-m2vsm_41be67bd-77e6-4b0b-bf54-9c3ffb228ea9/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:45 crc kubenswrapper[4992]: I1004 05:38:45.849731 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8699c8bbf-h2hwx_404db73d-c716-4d1a-808b-92b193c9b7cf/neutron-httpd/0.log" Oct 04 05:38:45 crc kubenswrapper[4992]: I1004 05:38:45.874880 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-8699c8bbf-h2hwx_404db73d-c716-4d1a-808b-92b193c9b7cf/neutron-api/0.log" Oct 04 05:38:46 crc kubenswrapper[4992]: I1004 05:38:46.148614 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-smv69_241b4766-f3d7-4953-845a-6a898cea420e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:46 crc kubenswrapper[4992]: I1004 05:38:46.748259 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23d3720b-cfd2-4513-93c0-6f9256dda97a/nova-api-log/0.log" Oct 04 05:38:46 crc kubenswrapper[4992]: I1004 05:38:46.899231 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f86bdf47-6f6d-4e6d-bcb4-c61cbd7744ee/nova-cell0-conductor-conductor/0.log" Oct 04 05:38:46 crc kubenswrapper[4992]: I1004 05:38:46.948149 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23d3720b-cfd2-4513-93c0-6f9256dda97a/nova-api-api/0.log" Oct 04 05:38:47 crc kubenswrapper[4992]: I1004 05:38:47.403453 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f291828a-08bf-400b-95d0-af8e6e7ecd10/nova-cell1-conductor-conductor/0.log" Oct 04 05:38:47 crc kubenswrapper[4992]: I1004 05:38:47.552072 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_1544458b-29c9-465f-adc2-f5ad257303fc/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 05:38:47 crc kubenswrapper[4992]: I1004 05:38:47.717965 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-8dcg2_a29da406-d169-4c98-ac72-45380f713377/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:47 crc kubenswrapper[4992]: I1004 05:38:47.946253 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f399940f-164b-4033-af0a-36359efe7229/nova-metadata-log/0.log" Oct 04 05:38:48 crc kubenswrapper[4992]: I1004 05:38:48.347443 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_bcf9657e-a56d-48fa-906a-aeaa6f3d3672/nova-scheduler-scheduler/0.log" Oct 04 05:38:48 crc kubenswrapper[4992]: I1004 05:38:48.982106 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ddfacb66-d69e-4113-993a-f64918d8fc0f/mysql-bootstrap/0.log" Oct 04 05:38:49 crc kubenswrapper[4992]: I1004 05:38:49.146471 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ddfacb66-d69e-4113-993a-f64918d8fc0f/mysql-bootstrap/0.log" Oct 04 05:38:49 crc kubenswrapper[4992]: I1004 05:38:49.211722 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ddfacb66-d69e-4113-993a-f64918d8fc0f/galera/0.log" Oct 04 05:38:49 crc kubenswrapper[4992]: I1004 05:38:49.384728 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f399940f-164b-4033-af0a-36359efe7229/nova-metadata-metadata/0.log" Oct 04 05:38:49 crc kubenswrapper[4992]: I1004 05:38:49.436642 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e7c0512f-849a-4178-b358-989265bddf4a/mysql-bootstrap/0.log" Oct 04 05:38:49 crc kubenswrapper[4992]: I1004 05:38:49.744794 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e7c0512f-849a-4178-b358-989265bddf4a/mysql-bootstrap/0.log" Oct 04 05:38:49 crc kubenswrapper[4992]: I1004 05:38:49.753110 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e7c0512f-849a-4178-b358-989265bddf4a/galera/0.log" Oct 04 05:38:49 crc kubenswrapper[4992]: I1004 05:38:49.931248 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_8a864935-07eb-48c7-a02b-a81ff050ce4b/openstackclient/0.log" Oct 04 05:38:50 crc kubenswrapper[4992]: I1004 05:38:50.035335 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-8m2tq_06640053-739f-4a61-84e0-c6ed4876a249/ovn-controller/0.log" Oct 04 05:38:50 crc kubenswrapper[4992]: I1004 05:38:50.262338 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-gbrkk_a564ac8c-42b9-48b3-9bd3-9785ab05b19d/openstack-network-exporter/0.log" Oct 04 05:38:50 crc kubenswrapper[4992]: I1004 05:38:50.320607 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:38:50 crc kubenswrapper[4992]: E1004 05:38:50.320815 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:38:50 crc kubenswrapper[4992]: I1004 05:38:50.457676 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qv6w9_68517c56-23d7-4353-ae99-eb10fee07a28/ovsdb-server-init/0.log" Oct 04 05:38:50 crc kubenswrapper[4992]: I1004 05:38:50.622063 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qv6w9_68517c56-23d7-4353-ae99-eb10fee07a28/ovs-vswitchd/0.log" Oct 04 05:38:50 crc kubenswrapper[4992]: I1004 05:38:50.628619 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qv6w9_68517c56-23d7-4353-ae99-eb10fee07a28/ovsdb-server-init/0.log" Oct 04 05:38:50 crc kubenswrapper[4992]: I1004 05:38:50.743619 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qv6w9_68517c56-23d7-4353-ae99-eb10fee07a28/ovsdb-server/0.log" Oct 04 05:38:50 crc kubenswrapper[4992]: I1004 05:38:50.901235 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-l6hlw_3185da64-244b-4f68-830f-f5d1cac08daa/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:51 crc kubenswrapper[4992]: I1004 05:38:51.075823 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9a31b3de-d7c5-4e1e-8ea8-fce11d130790/openstack-network-exporter/0.log" Oct 04 05:38:51 crc kubenswrapper[4992]: I1004 05:38:51.165805 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9a31b3de-d7c5-4e1e-8ea8-fce11d130790/ovn-northd/0.log" Oct 04 05:38:51 crc kubenswrapper[4992]: I1004 05:38:51.314289 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ae844ba9-4e34-4196-9ceb-17ecec8a6ce9/openstack-network-exporter/0.log" Oct 04 05:38:51 crc kubenswrapper[4992]: I1004 05:38:51.405412 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ae844ba9-4e34-4196-9ceb-17ecec8a6ce9/ovsdbserver-nb/0.log" Oct 04 05:38:51 crc kubenswrapper[4992]: I1004 05:38:51.513184 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f2338a74-bd09-4ef9-b79b-6f85817977ce/openstack-network-exporter/0.log" Oct 04 05:38:51 crc kubenswrapper[4992]: I1004 05:38:51.700434 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f2338a74-bd09-4ef9-b79b-6f85817977ce/ovsdbserver-sb/0.log" Oct 04 05:38:51 crc kubenswrapper[4992]: I1004 05:38:51.801043 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-689c5fd8b4-p4lxm_185b6e92-a7f2-4821-aed0-c732a0dc0e65/placement-api/0.log" Oct 04 05:38:51 crc kubenswrapper[4992]: I1004 05:38:51.983897 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-689c5fd8b4-p4lxm_185b6e92-a7f2-4821-aed0-c732a0dc0e65/placement-log/0.log" Oct 04 05:38:52 crc kubenswrapper[4992]: I1004 05:38:52.021537 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e3c63a70-cea7-4b0f-bf50-1f9dea7b1370/setup-container/0.log" Oct 04 05:38:52 crc kubenswrapper[4992]: I1004 05:38:52.323773 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e3c63a70-cea7-4b0f-bf50-1f9dea7b1370/rabbitmq/0.log" Oct 04 05:38:52 crc kubenswrapper[4992]: I1004 05:38:52.324647 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e3c63a70-cea7-4b0f-bf50-1f9dea7b1370/setup-container/0.log" Oct 04 05:38:52 crc kubenswrapper[4992]: I1004 05:38:52.482660 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_764fc1cd-ff45-46fe-8a26-e2dfdb02ed71/setup-container/0.log" Oct 04 05:38:52 crc kubenswrapper[4992]: I1004 05:38:52.816895 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_764fc1cd-ff45-46fe-8a26-e2dfdb02ed71/setup-container/0.log" Oct 04 05:38:52 crc kubenswrapper[4992]: I1004 05:38:52.874417 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_764fc1cd-ff45-46fe-8a26-e2dfdb02ed71/rabbitmq/0.log" Oct 04 05:38:53 crc kubenswrapper[4992]: I1004 05:38:53.060691 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-jhlrb_ef40937c-226e-49ce-83eb-944481fe7a82/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:53 crc kubenswrapper[4992]: I1004 05:38:53.460459 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-pmt2r_714c50bd-a6fa-44fb-ae36-f334c9bd9d68/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:53 crc kubenswrapper[4992]: I1004 05:38:53.460837 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-dbrd9_e910a377-6253-4eda-8850-6e1f2f25f99f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:53 crc kubenswrapper[4992]: I1004 05:38:53.676457 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-vtjm5_b57bbb97-a712-4ac2-841d-43b2053a009a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:53 crc kubenswrapper[4992]: I1004 05:38:53.779794 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-96vsc_04306ce6-821d-47e9-bfb1-765b7a067d44/ssh-known-hosts-edpm-deployment/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.026485 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-597c8b5b9f-s2g6d_ab81341f-2395-4a8d-a6c0-04ad2c2f970a/proxy-server/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.121010 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-597c8b5b9f-s2g6d_ab81341f-2395-4a8d-a6c0-04ad2c2f970a/proxy-httpd/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.259894 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-9hx4r_c9af92df-6216-469d-84e9-a09e3b8a648d/swift-ring-rebalance/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.459279 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/account-auditor/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.591607 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/account-reaper/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.691909 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/account-replicator/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.703036 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/account-server/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.851805 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/container-auditor/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.939169 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/container-server/0.log" Oct 04 05:38:54 crc kubenswrapper[4992]: I1004 05:38:54.955959 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/container-replicator/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.137080 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/container-updater/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.188951 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/object-auditor/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.216433 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/object-expirer/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.421547 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/object-replicator/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.428522 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/object-server/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.435781 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/object-updater/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.606807 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/rsync/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.609617 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_fa875baa-782c-469c-816f-7727c4885d28/swift-recon-cron/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.876829 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-fr8bs_8e39e2be-f116-452e-ac17-3b7337d80f5d/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:38:55 crc kubenswrapper[4992]: I1004 05:38:55.991370 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_dab26aab-5877-4a73-9433-7f76c31377dc/tempest-tests-tempest-tests-runner/0.log" Oct 04 05:38:56 crc kubenswrapper[4992]: I1004 05:38:56.147721 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_eb71ad30-b074-45c1-a11a-42eba81438f6/test-operator-logs-container/0.log" Oct 04 05:38:56 crc kubenswrapper[4992]: I1004 05:38:56.395329 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-ltx6f_9d3b0970-11af-4754-8ad0-beb5e97fd149/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 05:39:00 crc kubenswrapper[4992]: I1004 05:39:00.365727 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0a09ad3a-5a13-4ae7-a6d8-d9fb8d4fdde6/memcached/0.log" Oct 04 05:39:05 crc kubenswrapper[4992]: I1004 05:39:05.319666 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:39:05 crc kubenswrapper[4992]: E1004 05:39:05.320308 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:39:17 crc kubenswrapper[4992]: I1004 05:39:17.319307 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:39:17 crc kubenswrapper[4992]: E1004 05:39:17.320199 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:39:31 crc kubenswrapper[4992]: I1004 05:39:31.318893 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:39:31 crc kubenswrapper[4992]: E1004 05:39:31.319892 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:39:46 crc kubenswrapper[4992]: I1004 05:39:46.319584 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:39:46 crc kubenswrapper[4992]: E1004 05:39:46.320485 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:39:59 crc kubenswrapper[4992]: I1004 05:39:59.320652 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:39:59 crc kubenswrapper[4992]: E1004 05:39:59.321732 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:40:01 crc kubenswrapper[4992]: I1004 05:40:01.495715 4992 generic.go:334] "Generic (PLEG): container finished" podID="5f7b9547-55e2-4112-be8d-28191704a7c2" containerID="28ae4c509a0f46149029a591f04c1040344b5f1c58f1331dd8d0b02960c55680" exitCode=0 Oct 04 05:40:01 crc kubenswrapper[4992]: I1004 05:40:01.495763 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/crc-debug-wgkrl" event={"ID":"5f7b9547-55e2-4112-be8d-28191704a7c2","Type":"ContainerDied","Data":"28ae4c509a0f46149029a591f04c1040344b5f1c58f1331dd8d0b02960c55680"} Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.642943 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.672782 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-df5cv/crc-debug-wgkrl"] Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.679301 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-df5cv/crc-debug-wgkrl"] Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.841954 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f7b9547-55e2-4112-be8d-28191704a7c2-host\") pod \"5f7b9547-55e2-4112-be8d-28191704a7c2\" (UID: \"5f7b9547-55e2-4112-be8d-28191704a7c2\") " Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.842294 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74kz6\" (UniqueName: \"kubernetes.io/projected/5f7b9547-55e2-4112-be8d-28191704a7c2-kube-api-access-74kz6\") pod \"5f7b9547-55e2-4112-be8d-28191704a7c2\" (UID: \"5f7b9547-55e2-4112-be8d-28191704a7c2\") " Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.842042 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7b9547-55e2-4112-be8d-28191704a7c2-host" (OuterVolumeSpecName: "host") pod "5f7b9547-55e2-4112-be8d-28191704a7c2" (UID: "5f7b9547-55e2-4112-be8d-28191704a7c2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.842814 4992 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f7b9547-55e2-4112-be8d-28191704a7c2-host\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.848260 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f7b9547-55e2-4112-be8d-28191704a7c2-kube-api-access-74kz6" (OuterVolumeSpecName: "kube-api-access-74kz6") pod "5f7b9547-55e2-4112-be8d-28191704a7c2" (UID: "5f7b9547-55e2-4112-be8d-28191704a7c2"). InnerVolumeSpecName "kube-api-access-74kz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:40:02 crc kubenswrapper[4992]: I1004 05:40:02.944980 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74kz6\" (UniqueName: \"kubernetes.io/projected/5f7b9547-55e2-4112-be8d-28191704a7c2-kube-api-access-74kz6\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:03 crc kubenswrapper[4992]: I1004 05:40:03.518697 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dee7214a6b594f087941e93863c7c25e6416c19d96b33575e3a571937604a855" Oct 04 05:40:03 crc kubenswrapper[4992]: I1004 05:40:03.518746 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-wgkrl" Oct 04 05:40:03 crc kubenswrapper[4992]: I1004 05:40:03.853568 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-df5cv/crc-debug-hn9mv"] Oct 04 05:40:03 crc kubenswrapper[4992]: E1004 05:40:03.854134 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7b9547-55e2-4112-be8d-28191704a7c2" containerName="container-00" Oct 04 05:40:03 crc kubenswrapper[4992]: I1004 05:40:03.854158 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7b9547-55e2-4112-be8d-28191704a7c2" containerName="container-00" Oct 04 05:40:03 crc kubenswrapper[4992]: I1004 05:40:03.854493 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7b9547-55e2-4112-be8d-28191704a7c2" containerName="container-00" Oct 04 05:40:03 crc kubenswrapper[4992]: I1004 05:40:03.855484 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:03 crc kubenswrapper[4992]: I1004 05:40:03.962637 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n4zr\" (UniqueName: \"kubernetes.io/projected/3adba475-4e92-44cc-8d87-b8404b175e6d-kube-api-access-8n4zr\") pod \"crc-debug-hn9mv\" (UID: \"3adba475-4e92-44cc-8d87-b8404b175e6d\") " pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:03 crc kubenswrapper[4992]: I1004 05:40:03.963625 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3adba475-4e92-44cc-8d87-b8404b175e6d-host\") pod \"crc-debug-hn9mv\" (UID: \"3adba475-4e92-44cc-8d87-b8404b175e6d\") " pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:04 crc kubenswrapper[4992]: I1004 05:40:04.065972 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3adba475-4e92-44cc-8d87-b8404b175e6d-host\") pod \"crc-debug-hn9mv\" (UID: \"3adba475-4e92-44cc-8d87-b8404b175e6d\") " pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:04 crc kubenswrapper[4992]: I1004 05:40:04.066130 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3adba475-4e92-44cc-8d87-b8404b175e6d-host\") pod \"crc-debug-hn9mv\" (UID: \"3adba475-4e92-44cc-8d87-b8404b175e6d\") " pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:04 crc kubenswrapper[4992]: I1004 05:40:04.066355 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n4zr\" (UniqueName: \"kubernetes.io/projected/3adba475-4e92-44cc-8d87-b8404b175e6d-kube-api-access-8n4zr\") pod \"crc-debug-hn9mv\" (UID: \"3adba475-4e92-44cc-8d87-b8404b175e6d\") " pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:04 crc kubenswrapper[4992]: I1004 05:40:04.332390 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f7b9547-55e2-4112-be8d-28191704a7c2" path="/var/lib/kubelet/pods/5f7b9547-55e2-4112-be8d-28191704a7c2/volumes" Oct 04 05:40:04 crc kubenswrapper[4992]: I1004 05:40:04.422947 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n4zr\" (UniqueName: \"kubernetes.io/projected/3adba475-4e92-44cc-8d87-b8404b175e6d-kube-api-access-8n4zr\") pod \"crc-debug-hn9mv\" (UID: \"3adba475-4e92-44cc-8d87-b8404b175e6d\") " pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:04 crc kubenswrapper[4992]: I1004 05:40:04.472614 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:05 crc kubenswrapper[4992]: I1004 05:40:05.572666 4992 generic.go:334] "Generic (PLEG): container finished" podID="3adba475-4e92-44cc-8d87-b8404b175e6d" containerID="93be2a7a0a86af18aca124e8173151619ff1f6cedca4b6dfa5dea4ed802b8ac5" exitCode=0 Oct 04 05:40:05 crc kubenswrapper[4992]: I1004 05:40:05.572713 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/crc-debug-hn9mv" event={"ID":"3adba475-4e92-44cc-8d87-b8404b175e6d","Type":"ContainerDied","Data":"93be2a7a0a86af18aca124e8173151619ff1f6cedca4b6dfa5dea4ed802b8ac5"} Oct 04 05:40:05 crc kubenswrapper[4992]: I1004 05:40:05.572750 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/crc-debug-hn9mv" event={"ID":"3adba475-4e92-44cc-8d87-b8404b175e6d","Type":"ContainerStarted","Data":"2de07480bc187354068554d24adbf83b9ce02e2db7b0943a1f0a86cd68be3080"} Oct 04 05:40:06 crc kubenswrapper[4992]: I1004 05:40:06.701688 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:06 crc kubenswrapper[4992]: I1004 05:40:06.715297 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3adba475-4e92-44cc-8d87-b8404b175e6d-host\") pod \"3adba475-4e92-44cc-8d87-b8404b175e6d\" (UID: \"3adba475-4e92-44cc-8d87-b8404b175e6d\") " Oct 04 05:40:06 crc kubenswrapper[4992]: I1004 05:40:06.715352 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n4zr\" (UniqueName: \"kubernetes.io/projected/3adba475-4e92-44cc-8d87-b8404b175e6d-kube-api-access-8n4zr\") pod \"3adba475-4e92-44cc-8d87-b8404b175e6d\" (UID: \"3adba475-4e92-44cc-8d87-b8404b175e6d\") " Oct 04 05:40:06 crc kubenswrapper[4992]: I1004 05:40:06.715458 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3adba475-4e92-44cc-8d87-b8404b175e6d-host" (OuterVolumeSpecName: "host") pod "3adba475-4e92-44cc-8d87-b8404b175e6d" (UID: "3adba475-4e92-44cc-8d87-b8404b175e6d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 05:40:06 crc kubenswrapper[4992]: I1004 05:40:06.715756 4992 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3adba475-4e92-44cc-8d87-b8404b175e6d-host\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:06 crc kubenswrapper[4992]: I1004 05:40:06.724578 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3adba475-4e92-44cc-8d87-b8404b175e6d-kube-api-access-8n4zr" (OuterVolumeSpecName: "kube-api-access-8n4zr") pod "3adba475-4e92-44cc-8d87-b8404b175e6d" (UID: "3adba475-4e92-44cc-8d87-b8404b175e6d"). InnerVolumeSpecName "kube-api-access-8n4zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:40:06 crc kubenswrapper[4992]: I1004 05:40:06.816289 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n4zr\" (UniqueName: \"kubernetes.io/projected/3adba475-4e92-44cc-8d87-b8404b175e6d-kube-api-access-8n4zr\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:07 crc kubenswrapper[4992]: I1004 05:40:07.595574 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/crc-debug-hn9mv" event={"ID":"3adba475-4e92-44cc-8d87-b8404b175e6d","Type":"ContainerDied","Data":"2de07480bc187354068554d24adbf83b9ce02e2db7b0943a1f0a86cd68be3080"} Oct 04 05:40:07 crc kubenswrapper[4992]: I1004 05:40:07.595795 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2de07480bc187354068554d24adbf83b9ce02e2db7b0943a1f0a86cd68be3080" Oct 04 05:40:07 crc kubenswrapper[4992]: I1004 05:40:07.595848 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-hn9mv" Oct 04 05:40:11 crc kubenswrapper[4992]: I1004 05:40:11.318689 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:40:11 crc kubenswrapper[4992]: E1004 05:40:11.319475 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:40:12 crc kubenswrapper[4992]: I1004 05:40:12.068431 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-df5cv/crc-debug-hn9mv"] Oct 04 05:40:12 crc kubenswrapper[4992]: I1004 05:40:12.076418 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-df5cv/crc-debug-hn9mv"] Oct 04 05:40:12 crc kubenswrapper[4992]: I1004 05:40:12.626875 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3adba475-4e92-44cc-8d87-b8404b175e6d" path="/var/lib/kubelet/pods/3adba475-4e92-44cc-8d87-b8404b175e6d/volumes" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.227204 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-df5cv/crc-debug-vtpnr"] Oct 04 05:40:13 crc kubenswrapper[4992]: E1004 05:40:13.227663 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adba475-4e92-44cc-8d87-b8404b175e6d" containerName="container-00" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.227683 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adba475-4e92-44cc-8d87-b8404b175e6d" containerName="container-00" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.227881 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="3adba475-4e92-44cc-8d87-b8404b175e6d" containerName="container-00" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.228539 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.314253 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff7p4\" (UniqueName: \"kubernetes.io/projected/f49f8295-18a7-4a4f-950e-8e58de37071d-kube-api-access-ff7p4\") pod \"crc-debug-vtpnr\" (UID: \"f49f8295-18a7-4a4f-950e-8e58de37071d\") " pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.314836 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f49f8295-18a7-4a4f-950e-8e58de37071d-host\") pod \"crc-debug-vtpnr\" (UID: \"f49f8295-18a7-4a4f-950e-8e58de37071d\") " pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.416734 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f49f8295-18a7-4a4f-950e-8e58de37071d-host\") pod \"crc-debug-vtpnr\" (UID: \"f49f8295-18a7-4a4f-950e-8e58de37071d\") " pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.416907 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f49f8295-18a7-4a4f-950e-8e58de37071d-host\") pod \"crc-debug-vtpnr\" (UID: \"f49f8295-18a7-4a4f-950e-8e58de37071d\") " pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.416966 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff7p4\" (UniqueName: \"kubernetes.io/projected/f49f8295-18a7-4a4f-950e-8e58de37071d-kube-api-access-ff7p4\") pod \"crc-debug-vtpnr\" (UID: \"f49f8295-18a7-4a4f-950e-8e58de37071d\") " pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.444659 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff7p4\" (UniqueName: \"kubernetes.io/projected/f49f8295-18a7-4a4f-950e-8e58de37071d-kube-api-access-ff7p4\") pod \"crc-debug-vtpnr\" (UID: \"f49f8295-18a7-4a4f-950e-8e58de37071d\") " pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.548583 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:13 crc kubenswrapper[4992]: I1004 05:40:13.648846 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/crc-debug-vtpnr" event={"ID":"f49f8295-18a7-4a4f-950e-8e58de37071d","Type":"ContainerStarted","Data":"b8bc02ba02d7e4fe6fbfffbcc19f4c24f8beb1c37f243f5e6335c6f8f4de904e"} Oct 04 05:40:14 crc kubenswrapper[4992]: I1004 05:40:14.661141 4992 generic.go:334] "Generic (PLEG): container finished" podID="f49f8295-18a7-4a4f-950e-8e58de37071d" containerID="e67b7bcadaaa8e1358787e383136ae5de34566f100dcc38acb75119fbdd9672f" exitCode=0 Oct 04 05:40:14 crc kubenswrapper[4992]: I1004 05:40:14.661239 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/crc-debug-vtpnr" event={"ID":"f49f8295-18a7-4a4f-950e-8e58de37071d","Type":"ContainerDied","Data":"e67b7bcadaaa8e1358787e383136ae5de34566f100dcc38acb75119fbdd9672f"} Oct 04 05:40:14 crc kubenswrapper[4992]: I1004 05:40:14.705341 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-df5cv/crc-debug-vtpnr"] Oct 04 05:40:14 crc kubenswrapper[4992]: I1004 05:40:14.716690 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-df5cv/crc-debug-vtpnr"] Oct 04 05:40:15 crc kubenswrapper[4992]: I1004 05:40:15.778901 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:15 crc kubenswrapper[4992]: I1004 05:40:15.968556 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff7p4\" (UniqueName: \"kubernetes.io/projected/f49f8295-18a7-4a4f-950e-8e58de37071d-kube-api-access-ff7p4\") pod \"f49f8295-18a7-4a4f-950e-8e58de37071d\" (UID: \"f49f8295-18a7-4a4f-950e-8e58de37071d\") " Oct 04 05:40:15 crc kubenswrapper[4992]: I1004 05:40:15.968902 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f49f8295-18a7-4a4f-950e-8e58de37071d-host\") pod \"f49f8295-18a7-4a4f-950e-8e58de37071d\" (UID: \"f49f8295-18a7-4a4f-950e-8e58de37071d\") " Oct 04 05:40:15 crc kubenswrapper[4992]: I1004 05:40:15.969383 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f49f8295-18a7-4a4f-950e-8e58de37071d-host" (OuterVolumeSpecName: "host") pod "f49f8295-18a7-4a4f-950e-8e58de37071d" (UID: "f49f8295-18a7-4a4f-950e-8e58de37071d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 05:40:15 crc kubenswrapper[4992]: I1004 05:40:15.973792 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f49f8295-18a7-4a4f-950e-8e58de37071d-kube-api-access-ff7p4" (OuterVolumeSpecName: "kube-api-access-ff7p4") pod "f49f8295-18a7-4a4f-950e-8e58de37071d" (UID: "f49f8295-18a7-4a4f-950e-8e58de37071d"). InnerVolumeSpecName "kube-api-access-ff7p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:40:16 crc kubenswrapper[4992]: I1004 05:40:16.071467 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff7p4\" (UniqueName: \"kubernetes.io/projected/f49f8295-18a7-4a4f-950e-8e58de37071d-kube-api-access-ff7p4\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:16 crc kubenswrapper[4992]: I1004 05:40:16.071504 4992 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f49f8295-18a7-4a4f-950e-8e58de37071d-host\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:16 crc kubenswrapper[4992]: I1004 05:40:16.331556 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f49f8295-18a7-4a4f-950e-8e58de37071d" path="/var/lib/kubelet/pods/f49f8295-18a7-4a4f-950e-8e58de37071d/volumes" Oct 04 05:40:16 crc kubenswrapper[4992]: I1004 05:40:16.687035 4992 scope.go:117] "RemoveContainer" containerID="e67b7bcadaaa8e1358787e383136ae5de34566f100dcc38acb75119fbdd9672f" Oct 04 05:40:16 crc kubenswrapper[4992]: I1004 05:40:16.687228 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/crc-debug-vtpnr" Oct 04 05:40:16 crc kubenswrapper[4992]: I1004 05:40:16.879879 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w_c947400d-2e44-4d6d-b0c0-60814a3acdd6/util/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.017535 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w_c947400d-2e44-4d6d-b0c0-60814a3acdd6/util/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.073754 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w_c947400d-2e44-4d6d-b0c0-60814a3acdd6/pull/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.086341 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w_c947400d-2e44-4d6d-b0c0-60814a3acdd6/pull/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.293137 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w_c947400d-2e44-4d6d-b0c0-60814a3acdd6/pull/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.307905 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w_c947400d-2e44-4d6d-b0c0-60814a3acdd6/extract/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.315399 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_53462a13e994233cf65d23fed3d6d0af0e65872ec3f8f297c06d2bd5a9jts9w_c947400d-2e44-4d6d-b0c0-60814a3acdd6/util/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.502582 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-6gwsx_07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35/kube-rbac-proxy/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.562638 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-crk9q_a6b05a48-2a86-4afc-87c4-e6006f45b812/kube-rbac-proxy/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.604440 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5f7c849b98-crk9q_a6b05a48-2a86-4afc-87c4-e6006f45b812/manager/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.741854 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-6gwsx_07db2e1d-eb98-4ea9-bcda-a1ce07ecbd35/manager/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.785401 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-scgjp_b185c908-306b-4a41-b5f4-94e6024b050b/manager/0.log" Oct 04 05:40:17 crc kubenswrapper[4992]: I1004 05:40:17.834808 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-scgjp_b185c908-306b-4a41-b5f4-94e6024b050b/kube-rbac-proxy/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.072071 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-tctb4_5c53b372-1912-4b4a-9a73-8b0619d4c9c8/manager/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.081138 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5568b5d68-tctb4_5c53b372-1912-4b4a-9a73-8b0619d4c9c8/kube-rbac-proxy/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.210110 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-59d5d_bd358fd0-4f19-4420-9dee-7d8a2daafb83/kube-rbac-proxy/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.434080 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-8f58bc9db-59d5d_bd358fd0-4f19-4420-9dee-7d8a2daafb83/manager/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.486211 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-r42pj_7bb260b0-fb2e-4ad8-9ea9-198934b62e60/kube-rbac-proxy/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.613171 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54876c876f-r42pj_7bb260b0-fb2e-4ad8-9ea9-198934b62e60/manager/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.751608 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-czxxb_fdba06a8-d608-4111-a067-aedd5f292557/kube-rbac-proxy/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.844623 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-576fh_44ed3dc1-bf8a-4c86-90c2-71490a4ab882/kube-rbac-proxy/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.870645 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-czxxb_fdba06a8-d608-4111-a067-aedd5f292557/manager/0.log" Oct 04 05:40:18 crc kubenswrapper[4992]: I1004 05:40:18.979314 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-699b87f775-576fh_44ed3dc1-bf8a-4c86-90c2-71490a4ab882/manager/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.103157 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-5j2nq_4185f955-0390-47f4-88d3-18f5c7f9c4e5/kube-rbac-proxy/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.149107 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-655d88ccb9-5j2nq_4185f955-0390-47f4-88d3-18f5c7f9c4e5/manager/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.302432 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-4vlxm_411b554a-a210-4025-be0b-ba8d0a6e912b/kube-rbac-proxy/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.311799 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-4vlxm_411b554a-a210-4025-be0b-ba8d0a6e912b/manager/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.419858 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-t52mq_7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668/kube-rbac-proxy/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.553514 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-t52mq_7ad68004-a60a-4d5f-a4f8-fe6e7cfdc668/manager/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.574535 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-nbbkf_49e9ad0f-706f-4aa5-8e5d-347e29a408cc/kube-rbac-proxy/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.673027 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-nbbkf_49e9ad0f-706f-4aa5-8e5d-347e29a408cc/manager/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.849889 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-mj9np_94a4d61f-bf8a-436a-a267-f8b014d9418e/manager/0.log" Oct 04 05:40:19 crc kubenswrapper[4992]: I1004 05:40:19.857659 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-mj9np_94a4d61f-bf8a-436a-a267-f8b014d9418e/kube-rbac-proxy/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.000549 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-cnzlq_63cd2106-f93c-49af-88f3-a1598ab76e8b/kube-rbac-proxy/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.023239 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-cnzlq_63cd2106-f93c-49af-88f3-a1598ab76e8b/manager/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.063644 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2_70c5935b-fd73-4f67-928f-178131a1c7c5/kube-rbac-proxy/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.110713 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665c87lv2_70c5935b-fd73-4f67-928f-178131a1c7c5/manager/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.189557 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c6db87b77-kftlz_9736ec36-6280-4353-9e71-f64c8e6b5756/kube-rbac-proxy/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.369740 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6dc5776fd9-k9tqq_f3ae9a99-2aae-49d4-a881-8ff336de3c3f/kube-rbac-proxy/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.478958 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xp6bx_fd7bfac1-0f3c-4e68-9043-91eabbd07afb/registry-server/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.604206 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6dc5776fd9-k9tqq_f3ae9a99-2aae-49d4-a881-8ff336de3c3f/operator/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.777407 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-85jrh_75dc0bae-78b9-45f6-b8fd-f9ee9f307ace/kube-rbac-proxy/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.948417 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-vmkh5_463db7ad-b3a2-469e-9a74-cb170fb54b0d/kube-rbac-proxy/0.log" Oct 04 05:40:20 crc kubenswrapper[4992]: I1004 05:40:20.974351 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-579449c7d5-85jrh_75dc0bae-78b9-45f6-b8fd-f9ee9f307ace/manager/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.021023 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-vmkh5_463db7ad-b3a2-469e-9a74-cb170fb54b0d/manager/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.221578 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-6tfnt_ef4005a7-625d-49a8-a37a-25058eb5de15/operator/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.231216 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-dc9gg_d0d758eb-b8a5-4508-bdc2-8a58158d864f/kube-rbac-proxy/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.342207 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-dc9gg_d0d758eb-b8a5-4508-bdc2-8a58158d864f/manager/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.450510 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6c6db87b77-kftlz_9736ec36-6280-4353-9e71-f64c8e6b5756/manager/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.494910 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-phbgj_892ab60b-1964-4943-9238-eedb8a72b3e4/manager/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.585703 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-phbgj_892ab60b-1964-4943-9238-eedb8a72b3e4/kube-rbac-proxy/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.605700 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-kf7nd_ca9f0c2e-15e3-404f-83b6-a9a438ca88c9/kube-rbac-proxy/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.643496 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-kf7nd_ca9f0c2e-15e3-404f-83b6-a9a438ca88c9/manager/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.777714 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-mdvtr_cbcb9dfe-fcdc-4adf-b65c-346aded84b3f/kube-rbac-proxy/0.log" Oct 04 05:40:21 crc kubenswrapper[4992]: I1004 05:40:21.790186 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-mdvtr_cbcb9dfe-fcdc-4adf-b65c-346aded84b3f/manager/0.log" Oct 04 05:40:26 crc kubenswrapper[4992]: I1004 05:40:26.319645 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:40:26 crc kubenswrapper[4992]: E1004 05:40:26.320447 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:40:38 crc kubenswrapper[4992]: I1004 05:40:38.142340 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4f9mb_9d143445-9a26-4203-995d-82238aa12f54/control-plane-machine-set-operator/0.log" Oct 04 05:40:38 crc kubenswrapper[4992]: I1004 05:40:38.320987 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-pbx8v_b240811c-d6e5-4a4f-932a-fdf41ad86a98/machine-api-operator/0.log" Oct 04 05:40:38 crc kubenswrapper[4992]: I1004 05:40:38.326231 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-pbx8v_b240811c-d6e5-4a4f-932a-fdf41ad86a98/kube-rbac-proxy/0.log" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.319666 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:40:40 crc kubenswrapper[4992]: E1004 05:40:40.320216 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.738632 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d2rqz"] Oct 04 05:40:40 crc kubenswrapper[4992]: E1004 05:40:40.739303 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f49f8295-18a7-4a4f-950e-8e58de37071d" containerName="container-00" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.739417 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="f49f8295-18a7-4a4f-950e-8e58de37071d" containerName="container-00" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.739737 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="f49f8295-18a7-4a4f-950e-8e58de37071d" containerName="container-00" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.741448 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.751352 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d2rqz"] Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.883786 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs2p5\" (UniqueName: \"kubernetes.io/projected/08b6b869-f803-4d19-903c-47e80e2c958f-kube-api-access-qs2p5\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.884003 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-catalog-content\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.884035 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-utilities\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.985836 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-catalog-content\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.985893 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-utilities\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.986017 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs2p5\" (UniqueName: \"kubernetes.io/projected/08b6b869-f803-4d19-903c-47e80e2c958f-kube-api-access-qs2p5\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.986568 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-utilities\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:40 crc kubenswrapper[4992]: I1004 05:40:40.986650 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-catalog-content\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:41 crc kubenswrapper[4992]: I1004 05:40:41.006503 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs2p5\" (UniqueName: \"kubernetes.io/projected/08b6b869-f803-4d19-903c-47e80e2c958f-kube-api-access-qs2p5\") pod \"community-operators-d2rqz\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:41 crc kubenswrapper[4992]: I1004 05:40:41.078433 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:41 crc kubenswrapper[4992]: I1004 05:40:41.537846 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d2rqz"] Oct 04 05:40:41 crc kubenswrapper[4992]: I1004 05:40:41.963309 4992 generic.go:334] "Generic (PLEG): container finished" podID="08b6b869-f803-4d19-903c-47e80e2c958f" containerID="f5932de8fef91592c312ab73d4fb641cac92e8fbcfe8fda01b801f3aeb09b7fb" exitCode=0 Oct 04 05:40:41 crc kubenswrapper[4992]: I1004 05:40:41.963354 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2rqz" event={"ID":"08b6b869-f803-4d19-903c-47e80e2c958f","Type":"ContainerDied","Data":"f5932de8fef91592c312ab73d4fb641cac92e8fbcfe8fda01b801f3aeb09b7fb"} Oct 04 05:40:41 crc kubenswrapper[4992]: I1004 05:40:41.963645 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2rqz" event={"ID":"08b6b869-f803-4d19-903c-47e80e2c958f","Type":"ContainerStarted","Data":"18a9850508b7bdcc1a01cfb100182b441a6b816f2f92f1c66f5189931f786524"} Oct 04 05:40:42 crc kubenswrapper[4992]: I1004 05:40:42.975056 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2rqz" event={"ID":"08b6b869-f803-4d19-903c-47e80e2c958f","Type":"ContainerStarted","Data":"db3f8c539a87cc3e7adc8f0ce7ff070024b4e17abdf65e4038f3c6acbf52ac07"} Oct 04 05:40:43 crc kubenswrapper[4992]: I1004 05:40:43.987097 4992 generic.go:334] "Generic (PLEG): container finished" podID="08b6b869-f803-4d19-903c-47e80e2c958f" containerID="db3f8c539a87cc3e7adc8f0ce7ff070024b4e17abdf65e4038f3c6acbf52ac07" exitCode=0 Oct 04 05:40:43 crc kubenswrapper[4992]: I1004 05:40:43.987450 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2rqz" event={"ID":"08b6b869-f803-4d19-903c-47e80e2c958f","Type":"ContainerDied","Data":"db3f8c539a87cc3e7adc8f0ce7ff070024b4e17abdf65e4038f3c6acbf52ac07"} Oct 04 05:40:46 crc kubenswrapper[4992]: I1004 05:40:46.007870 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2rqz" event={"ID":"08b6b869-f803-4d19-903c-47e80e2c958f","Type":"ContainerStarted","Data":"faa63ef8d4d0c56b83fd497d426297ae44e6e277f5c2613ee27bf6d87dbae25e"} Oct 04 05:40:46 crc kubenswrapper[4992]: I1004 05:40:46.033587 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d2rqz" podStartSLOduration=3.567653809 podStartE2EDuration="6.033568149s" podCreationTimestamp="2025-10-04 05:40:40 +0000 UTC" firstStartedPulling="2025-10-04 05:40:41.965244691 +0000 UTC m=+4075.812920159" lastFinishedPulling="2025-10-04 05:40:44.431159021 +0000 UTC m=+4078.278834499" observedRunningTime="2025-10-04 05:40:46.023717916 +0000 UTC m=+4079.871393404" watchObservedRunningTime="2025-10-04 05:40:46.033568149 +0000 UTC m=+4079.881243617" Oct 04 05:40:50 crc kubenswrapper[4992]: I1004 05:40:50.723796 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7tqhx_059e879c-9cec-4e22-8702-8c5f93cffd5d/cert-manager-controller/0.log" Oct 04 05:40:50 crc kubenswrapper[4992]: I1004 05:40:50.880793 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-4gjdf_eefe44db-fa85-476c-9900-7c1e3d46a63f/cert-manager-cainjector/0.log" Oct 04 05:40:50 crc kubenswrapper[4992]: I1004 05:40:50.931530 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-5lblx_911792e6-61de-4469-9ced-16ffca1da0e7/cert-manager-webhook/0.log" Oct 04 05:40:51 crc kubenswrapper[4992]: I1004 05:40:51.079214 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:51 crc kubenswrapper[4992]: I1004 05:40:51.079575 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:51 crc kubenswrapper[4992]: I1004 05:40:51.133162 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:52 crc kubenswrapper[4992]: I1004 05:40:52.122039 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:52 crc kubenswrapper[4992]: I1004 05:40:52.170050 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d2rqz"] Oct 04 05:40:54 crc kubenswrapper[4992]: I1004 05:40:54.080626 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d2rqz" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" containerName="registry-server" containerID="cri-o://faa63ef8d4d0c56b83fd497d426297ae44e6e277f5c2613ee27bf6d87dbae25e" gracePeriod=2 Oct 04 05:40:55 crc kubenswrapper[4992]: I1004 05:40:55.092772 4992 generic.go:334] "Generic (PLEG): container finished" podID="08b6b869-f803-4d19-903c-47e80e2c958f" containerID="faa63ef8d4d0c56b83fd497d426297ae44e6e277f5c2613ee27bf6d87dbae25e" exitCode=0 Oct 04 05:40:55 crc kubenswrapper[4992]: I1004 05:40:55.093008 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2rqz" event={"ID":"08b6b869-f803-4d19-903c-47e80e2c958f","Type":"ContainerDied","Data":"faa63ef8d4d0c56b83fd497d426297ae44e6e277f5c2613ee27bf6d87dbae25e"} Oct 04 05:40:55 crc kubenswrapper[4992]: I1004 05:40:55.320463 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:40:55 crc kubenswrapper[4992]: E1004 05:40:55.320891 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.333189 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.503148 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs2p5\" (UniqueName: \"kubernetes.io/projected/08b6b869-f803-4d19-903c-47e80e2c958f-kube-api-access-qs2p5\") pod \"08b6b869-f803-4d19-903c-47e80e2c958f\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.503592 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-utilities\") pod \"08b6b869-f803-4d19-903c-47e80e2c958f\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.503674 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-catalog-content\") pod \"08b6b869-f803-4d19-903c-47e80e2c958f\" (UID: \"08b6b869-f803-4d19-903c-47e80e2c958f\") " Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.504612 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-utilities" (OuterVolumeSpecName: "utilities") pod "08b6b869-f803-4d19-903c-47e80e2c958f" (UID: "08b6b869-f803-4d19-903c-47e80e2c958f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.510202 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b6b869-f803-4d19-903c-47e80e2c958f-kube-api-access-qs2p5" (OuterVolumeSpecName: "kube-api-access-qs2p5") pod "08b6b869-f803-4d19-903c-47e80e2c958f" (UID: "08b6b869-f803-4d19-903c-47e80e2c958f"). InnerVolumeSpecName "kube-api-access-qs2p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.554146 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08b6b869-f803-4d19-903c-47e80e2c958f" (UID: "08b6b869-f803-4d19-903c-47e80e2c958f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.605717 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs2p5\" (UniqueName: \"kubernetes.io/projected/08b6b869-f803-4d19-903c-47e80e2c958f-kube-api-access-qs2p5\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.605753 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:56 crc kubenswrapper[4992]: I1004 05:40:56.605763 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6b869-f803-4d19-903c-47e80e2c958f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:40:57 crc kubenswrapper[4992]: I1004 05:40:57.123217 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2rqz" event={"ID":"08b6b869-f803-4d19-903c-47e80e2c958f","Type":"ContainerDied","Data":"18a9850508b7bdcc1a01cfb100182b441a6b816f2f92f1c66f5189931f786524"} Oct 04 05:40:57 crc kubenswrapper[4992]: I1004 05:40:57.123474 4992 scope.go:117] "RemoveContainer" containerID="faa63ef8d4d0c56b83fd497d426297ae44e6e277f5c2613ee27bf6d87dbae25e" Oct 04 05:40:57 crc kubenswrapper[4992]: I1004 05:40:57.123618 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2rqz" Oct 04 05:40:57 crc kubenswrapper[4992]: I1004 05:40:57.153045 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d2rqz"] Oct 04 05:40:57 crc kubenswrapper[4992]: I1004 05:40:57.161019 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d2rqz"] Oct 04 05:40:57 crc kubenswrapper[4992]: I1004 05:40:57.168586 4992 scope.go:117] "RemoveContainer" containerID="db3f8c539a87cc3e7adc8f0ce7ff070024b4e17abdf65e4038f3c6acbf52ac07" Oct 04 05:40:57 crc kubenswrapper[4992]: I1004 05:40:57.196107 4992 scope.go:117] "RemoveContainer" containerID="f5932de8fef91592c312ab73d4fb641cac92e8fbcfe8fda01b801f3aeb09b7fb" Oct 04 05:40:58 crc kubenswrapper[4992]: I1004 05:40:58.334417 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" path="/var/lib/kubelet/pods/08b6b869-f803-4d19-903c-47e80e2c958f/volumes" Oct 04 05:41:03 crc kubenswrapper[4992]: I1004 05:41:03.515504 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-s2crp_3bba2412-53af-4a50-aa28-439c21630dd6/nmstate-console-plugin/0.log" Oct 04 05:41:03 crc kubenswrapper[4992]: I1004 05:41:03.710850 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-xn8rv_e65ecf64-2e05-4ce7-a67a-ec85b8c39d67/nmstate-handler/0.log" Oct 04 05:41:03 crc kubenswrapper[4992]: I1004 05:41:03.765061 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-qncn9_c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0/kube-rbac-proxy/0.log" Oct 04 05:41:03 crc kubenswrapper[4992]: I1004 05:41:03.783310 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-qncn9_c1f8f275-8e3a-4ed4-ae06-7a14cd5b93f0/nmstate-metrics/0.log" Oct 04 05:41:03 crc kubenswrapper[4992]: I1004 05:41:03.894627 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-kr2c6_dd3e49fa-dfd8-4fd8-bd42-4bf8a3f14d7a/nmstate-operator/0.log" Oct 04 05:41:03 crc kubenswrapper[4992]: I1004 05:41:03.967746 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-kvc8z_00d2e67f-7b6a-43dd-afd0-500f16e5ae3c/nmstate-webhook/0.log" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.137300 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tmkd9"] Oct 04 05:41:04 crc kubenswrapper[4992]: E1004 05:41:04.138440 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" containerName="registry-server" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.138461 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" containerName="registry-server" Oct 04 05:41:04 crc kubenswrapper[4992]: E1004 05:41:04.138489 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" containerName="extract-utilities" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.138503 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" containerName="extract-utilities" Oct 04 05:41:04 crc kubenswrapper[4992]: E1004 05:41:04.138546 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" containerName="extract-content" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.138555 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" containerName="extract-content" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.139136 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b6b869-f803-4d19-903c-47e80e2c958f" containerName="registry-server" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.150575 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.196772 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tmkd9"] Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.250988 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-catalog-content\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.251047 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttr27\" (UniqueName: \"kubernetes.io/projected/b24db047-390c-4acb-9770-536264556a07-kube-api-access-ttr27\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.251116 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-utilities\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.352226 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-utilities\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.352373 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-catalog-content\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.352418 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttr27\" (UniqueName: \"kubernetes.io/projected/b24db047-390c-4acb-9770-536264556a07-kube-api-access-ttr27\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.352830 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-catalog-content\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.352839 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-utilities\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.371154 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttr27\" (UniqueName: \"kubernetes.io/projected/b24db047-390c-4acb-9770-536264556a07-kube-api-access-ttr27\") pod \"certified-operators-tmkd9\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:04 crc kubenswrapper[4992]: I1004 05:41:04.480560 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:05 crc kubenswrapper[4992]: I1004 05:41:05.402130 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tmkd9"] Oct 04 05:41:06 crc kubenswrapper[4992]: I1004 05:41:06.215718 4992 generic.go:334] "Generic (PLEG): container finished" podID="b24db047-390c-4acb-9770-536264556a07" containerID="c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2" exitCode=0 Oct 04 05:41:06 crc kubenswrapper[4992]: I1004 05:41:06.215976 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tmkd9" event={"ID":"b24db047-390c-4acb-9770-536264556a07","Type":"ContainerDied","Data":"c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2"} Oct 04 05:41:06 crc kubenswrapper[4992]: I1004 05:41:06.216000 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tmkd9" event={"ID":"b24db047-390c-4acb-9770-536264556a07","Type":"ContainerStarted","Data":"55c35c6d6f2c036760f31ed76ee9b36cdce663faded3e76a290ac17f2d162c3c"} Oct 04 05:41:07 crc kubenswrapper[4992]: I1004 05:41:07.228060 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tmkd9" event={"ID":"b24db047-390c-4acb-9770-536264556a07","Type":"ContainerStarted","Data":"6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309"} Oct 04 05:41:08 crc kubenswrapper[4992]: I1004 05:41:08.239433 4992 generic.go:334] "Generic (PLEG): container finished" podID="b24db047-390c-4acb-9770-536264556a07" containerID="6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309" exitCode=0 Oct 04 05:41:08 crc kubenswrapper[4992]: I1004 05:41:08.239573 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tmkd9" event={"ID":"b24db047-390c-4acb-9770-536264556a07","Type":"ContainerDied","Data":"6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309"} Oct 04 05:41:09 crc kubenswrapper[4992]: I1004 05:41:09.320841 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:41:09 crc kubenswrapper[4992]: E1004 05:41:09.321392 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:41:10 crc kubenswrapper[4992]: I1004 05:41:10.263778 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tmkd9" event={"ID":"b24db047-390c-4acb-9770-536264556a07","Type":"ContainerStarted","Data":"9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2"} Oct 04 05:41:10 crc kubenswrapper[4992]: I1004 05:41:10.284241 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tmkd9" podStartSLOduration=3.431046458 podStartE2EDuration="6.284224457s" podCreationTimestamp="2025-10-04 05:41:04 +0000 UTC" firstStartedPulling="2025-10-04 05:41:06.217814269 +0000 UTC m=+4100.065489737" lastFinishedPulling="2025-10-04 05:41:09.070992268 +0000 UTC m=+4102.918667736" observedRunningTime="2025-10-04 05:41:10.281253907 +0000 UTC m=+4104.128929375" watchObservedRunningTime="2025-10-04 05:41:10.284224457 +0000 UTC m=+4104.131899925" Oct 04 05:41:14 crc kubenswrapper[4992]: I1004 05:41:14.481216 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:14 crc kubenswrapper[4992]: I1004 05:41:14.483168 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:14 crc kubenswrapper[4992]: I1004 05:41:14.565501 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:15 crc kubenswrapper[4992]: I1004 05:41:15.370276 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:15 crc kubenswrapper[4992]: I1004 05:41:15.423025 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tmkd9"] Oct 04 05:41:17 crc kubenswrapper[4992]: I1004 05:41:17.331576 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tmkd9" podUID="b24db047-390c-4acb-9770-536264556a07" containerName="registry-server" containerID="cri-o://9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2" gracePeriod=2 Oct 04 05:41:17 crc kubenswrapper[4992]: I1004 05:41:17.917639 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.010756 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-utilities\") pod \"b24db047-390c-4acb-9770-536264556a07\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.010958 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttr27\" (UniqueName: \"kubernetes.io/projected/b24db047-390c-4acb-9770-536264556a07-kube-api-access-ttr27\") pod \"b24db047-390c-4acb-9770-536264556a07\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.011022 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-catalog-content\") pod \"b24db047-390c-4acb-9770-536264556a07\" (UID: \"b24db047-390c-4acb-9770-536264556a07\") " Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.011707 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-utilities" (OuterVolumeSpecName: "utilities") pod "b24db047-390c-4acb-9770-536264556a07" (UID: "b24db047-390c-4acb-9770-536264556a07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.029546 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24db047-390c-4acb-9770-536264556a07-kube-api-access-ttr27" (OuterVolumeSpecName: "kube-api-access-ttr27") pod "b24db047-390c-4acb-9770-536264556a07" (UID: "b24db047-390c-4acb-9770-536264556a07"). InnerVolumeSpecName "kube-api-access-ttr27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.060188 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b24db047-390c-4acb-9770-536264556a07" (UID: "b24db047-390c-4acb-9770-536264556a07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.113928 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.113975 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttr27\" (UniqueName: \"kubernetes.io/projected/b24db047-390c-4acb-9770-536264556a07-kube-api-access-ttr27\") on node \"crc\" DevicePath \"\"" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.113989 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b24db047-390c-4acb-9770-536264556a07-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.341396 4992 generic.go:334] "Generic (PLEG): container finished" podID="b24db047-390c-4acb-9770-536264556a07" containerID="9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2" exitCode=0 Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.341609 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tmkd9" event={"ID":"b24db047-390c-4acb-9770-536264556a07","Type":"ContainerDied","Data":"9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2"} Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.341633 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tmkd9" event={"ID":"b24db047-390c-4acb-9770-536264556a07","Type":"ContainerDied","Data":"55c35c6d6f2c036760f31ed76ee9b36cdce663faded3e76a290ac17f2d162c3c"} Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.341650 4992 scope.go:117] "RemoveContainer" containerID="9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.341774 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tmkd9" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.374105 4992 scope.go:117] "RemoveContainer" containerID="6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.374246 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tmkd9"] Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.387741 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tmkd9"] Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.408827 4992 scope.go:117] "RemoveContainer" containerID="c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.435161 4992 scope.go:117] "RemoveContainer" containerID="9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2" Oct 04 05:41:18 crc kubenswrapper[4992]: E1004 05:41:18.435641 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2\": container with ID starting with 9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2 not found: ID does not exist" containerID="9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.435674 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2"} err="failed to get container status \"9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2\": rpc error: code = NotFound desc = could not find container \"9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2\": container with ID starting with 9bf512d0266057215c23ac9cb53782fc972fba9cdee7b20f5c1ce413bf5fa6a2 not found: ID does not exist" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.435696 4992 scope.go:117] "RemoveContainer" containerID="6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309" Oct 04 05:41:18 crc kubenswrapper[4992]: E1004 05:41:18.436432 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309\": container with ID starting with 6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309 not found: ID does not exist" containerID="6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.436491 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309"} err="failed to get container status \"6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309\": rpc error: code = NotFound desc = could not find container \"6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309\": container with ID starting with 6175ec4405255d48a90f49fa24f9b9f748bb4f320cf09f436fa75815193b0309 not found: ID does not exist" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.436506 4992 scope.go:117] "RemoveContainer" containerID="c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2" Oct 04 05:41:18 crc kubenswrapper[4992]: E1004 05:41:18.436956 4992 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2\": container with ID starting with c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2 not found: ID does not exist" containerID="c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.436975 4992 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2"} err="failed to get container status \"c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2\": rpc error: code = NotFound desc = could not find container \"c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2\": container with ID starting with c775356505cd510455fd914b2ba3228dcd67eb5718d05e877b86d805dcd5c3d2 not found: ID does not exist" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.596128 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-lkqgp_3f777561-a01c-4e77-98f4-eb7b15f56eca/kube-rbac-proxy/0.log" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.619709 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-lkqgp_3f777561-a01c-4e77-98f4-eb7b15f56eca/controller/0.log" Oct 04 05:41:18 crc kubenswrapper[4992]: I1004 05:41:18.803630 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-frr-files/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.010014 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-reloader/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.015040 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-frr-files/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.020019 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-metrics/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.021573 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-reloader/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.258784 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-metrics/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.266323 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-reloader/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.268268 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-frr-files/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.273128 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-metrics/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.421940 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-metrics/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.474317 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-reloader/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.475338 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/controller/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.485819 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/cp-frr-files/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.682654 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/kube-rbac-proxy-frr/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.709592 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/frr-metrics/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.757009 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/kube-rbac-proxy/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.872169 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/reloader/0.log" Oct 04 05:41:19 crc kubenswrapper[4992]: I1004 05:41:19.939339 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-8tns7_64b51eb5-0b5d-4cc2-ada8-889a67608f62/frr-k8s-webhook-server/0.log" Oct 04 05:41:20 crc kubenswrapper[4992]: I1004 05:41:20.134682 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5fb67947c7-9kpxd_8fb02cec-1648-47c4-a2af-034fa5f204ca/manager/0.log" Oct 04 05:41:20 crc kubenswrapper[4992]: I1004 05:41:20.280380 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-8b8bf6765-q6l5d_f5ef1e00-0f56-41fb-8961-1a0638414444/webhook-server/0.log" Oct 04 05:41:20 crc kubenswrapper[4992]: I1004 05:41:20.330294 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24db047-390c-4acb-9770-536264556a07" path="/var/lib/kubelet/pods/b24db047-390c-4acb-9770-536264556a07/volumes" Oct 04 05:41:20 crc kubenswrapper[4992]: I1004 05:41:20.415583 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nbkqc_17e67a6f-46b1-434e-a422-21f96bf870ea/kube-rbac-proxy/0.log" Oct 04 05:41:20 crc kubenswrapper[4992]: I1004 05:41:20.988344 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-nbkqc_17e67a6f-46b1-434e-a422-21f96bf870ea/speaker/0.log" Oct 04 05:41:21 crc kubenswrapper[4992]: I1004 05:41:21.065588 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lb5zk_6e15c26c-ca04-4b97-bb73-c1dbdb9d23b1/frr/0.log" Oct 04 05:41:23 crc kubenswrapper[4992]: I1004 05:41:23.319275 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:41:23 crc kubenswrapper[4992]: E1004 05:41:23.319901 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:41:33 crc kubenswrapper[4992]: I1004 05:41:33.763738 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh_ca142652-0451-4c15-bd36-d2fc14a916ad/util/0.log" Oct 04 05:41:33 crc kubenswrapper[4992]: I1004 05:41:33.949178 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh_ca142652-0451-4c15-bd36-d2fc14a916ad/util/0.log" Oct 04 05:41:33 crc kubenswrapper[4992]: I1004 05:41:33.978801 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh_ca142652-0451-4c15-bd36-d2fc14a916ad/pull/0.log" Oct 04 05:41:33 crc kubenswrapper[4992]: I1004 05:41:33.984337 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh_ca142652-0451-4c15-bd36-d2fc14a916ad/pull/0.log" Oct 04 05:41:34 crc kubenswrapper[4992]: I1004 05:41:34.334076 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:41:34 crc kubenswrapper[4992]: E1004 05:41:34.334425 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:41:34 crc kubenswrapper[4992]: I1004 05:41:34.678638 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh_ca142652-0451-4c15-bd36-d2fc14a916ad/extract/0.log" Oct 04 05:41:34 crc kubenswrapper[4992]: I1004 05:41:34.680907 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh_ca142652-0451-4c15-bd36-d2fc14a916ad/util/0.log" Oct 04 05:41:34 crc kubenswrapper[4992]: I1004 05:41:34.683896 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d29cvzh_ca142652-0451-4c15-bd36-d2fc14a916ad/pull/0.log" Oct 04 05:41:34 crc kubenswrapper[4992]: I1004 05:41:34.860109 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lp25k_166adf65-fdf7-422e-b3a8-43ff22d793a5/extract-utilities/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.011336 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lp25k_166adf65-fdf7-422e-b3a8-43ff22d793a5/extract-content/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.013072 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lp25k_166adf65-fdf7-422e-b3a8-43ff22d793a5/extract-content/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.033197 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lp25k_166adf65-fdf7-422e-b3a8-43ff22d793a5/extract-utilities/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.201230 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lp25k_166adf65-fdf7-422e-b3a8-43ff22d793a5/extract-utilities/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.249020 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lp25k_166adf65-fdf7-422e-b3a8-43ff22d793a5/extract-content/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.447395 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn7c7_aa70c135-0efc-4f2e-bf72-25b6055cda88/extract-utilities/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.587913 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn7c7_aa70c135-0efc-4f2e-bf72-25b6055cda88/extract-utilities/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.600586 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn7c7_aa70c135-0efc-4f2e-bf72-25b6055cda88/extract-content/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.692141 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn7c7_aa70c135-0efc-4f2e-bf72-25b6055cda88/extract-content/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.868455 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn7c7_aa70c135-0efc-4f2e-bf72-25b6055cda88/extract-content/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.882715 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lp25k_166adf65-fdf7-422e-b3a8-43ff22d793a5/registry-server/0.log" Oct 04 05:41:35 crc kubenswrapper[4992]: I1004 05:41:35.904904 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn7c7_aa70c135-0efc-4f2e-bf72-25b6055cda88/extract-utilities/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.117341 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pn7c7_aa70c135-0efc-4f2e-bf72-25b6055cda88/registry-server/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.288942 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn_a5658ee6-1e89-477f-a4d3-f871bd8ca728/util/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.487518 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn_a5658ee6-1e89-477f-a4d3-f871bd8ca728/pull/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.506046 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn_a5658ee6-1e89-477f-a4d3-f871bd8ca728/pull/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.512874 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn_a5658ee6-1e89-477f-a4d3-f871bd8ca728/util/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.695672 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn_a5658ee6-1e89-477f-a4d3-f871bd8ca728/util/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.697190 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn_a5658ee6-1e89-477f-a4d3-f871bd8ca728/pull/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.759023 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cph2sn_a5658ee6-1e89-477f-a4d3-f871bd8ca728/extract/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.879313 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ztwc8_c8584f53-da6d-4606-8ac5-557d0fa24828/marketplace-operator/0.log" Oct 04 05:41:36 crc kubenswrapper[4992]: I1004 05:41:36.966418 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2k7x_36aa6d66-c182-47cb-8899-588cbe55f588/extract-utilities/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.120513 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2k7x_36aa6d66-c182-47cb-8899-588cbe55f588/extract-content/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.120828 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2k7x_36aa6d66-c182-47cb-8899-588cbe55f588/extract-utilities/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.151799 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2k7x_36aa6d66-c182-47cb-8899-588cbe55f588/extract-content/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.279677 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2k7x_36aa6d66-c182-47cb-8899-588cbe55f588/extract-utilities/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.297697 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2k7x_36aa6d66-c182-47cb-8899-588cbe55f588/extract-content/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.361248 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mq7hd_c0186a4f-5104-4beb-a51e-5aec49ed8f0c/extract-utilities/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.522475 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2k7x_36aa6d66-c182-47cb-8899-588cbe55f588/registry-server/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.536336 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mq7hd_c0186a4f-5104-4beb-a51e-5aec49ed8f0c/extract-content/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.560165 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mq7hd_c0186a4f-5104-4beb-a51e-5aec49ed8f0c/extract-utilities/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.595174 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mq7hd_c0186a4f-5104-4beb-a51e-5aec49ed8f0c/extract-content/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.742959 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mq7hd_c0186a4f-5104-4beb-a51e-5aec49ed8f0c/extract-utilities/0.log" Oct 04 05:41:37 crc kubenswrapper[4992]: I1004 05:41:37.778994 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mq7hd_c0186a4f-5104-4beb-a51e-5aec49ed8f0c/extract-content/0.log" Oct 04 05:41:38 crc kubenswrapper[4992]: I1004 05:41:38.320965 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mq7hd_c0186a4f-5104-4beb-a51e-5aec49ed8f0c/registry-server/0.log" Oct 04 05:41:47 crc kubenswrapper[4992]: I1004 05:41:47.319266 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:41:47 crc kubenswrapper[4992]: E1004 05:41:47.320187 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:42:00 crc kubenswrapper[4992]: I1004 05:42:00.319389 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:42:00 crc kubenswrapper[4992]: E1004 05:42:00.320114 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:42:15 crc kubenswrapper[4992]: I1004 05:42:15.319664 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:42:15 crc kubenswrapper[4992]: E1004 05:42:15.320957 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:42:29 crc kubenswrapper[4992]: I1004 05:42:29.318812 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:42:29 crc kubenswrapper[4992]: E1004 05:42:29.319849 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:42:40 crc kubenswrapper[4992]: I1004 05:42:40.319812 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:42:40 crc kubenswrapper[4992]: E1004 05:42:40.320481 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:42:52 crc kubenswrapper[4992]: I1004 05:42:52.320095 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:42:52 crc kubenswrapper[4992]: E1004 05:42:52.321230 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:43:06 crc kubenswrapper[4992]: I1004 05:43:06.319670 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:43:06 crc kubenswrapper[4992]: E1004 05:43:06.321142 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:43:17 crc kubenswrapper[4992]: I1004 05:43:17.319003 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:43:17 crc kubenswrapper[4992]: E1004 05:43:17.319481 4992 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p5l9q_openshift-machine-config-operator(d6d4bf54-7f00-4c33-830c-8368dd9678df)\"" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" Oct 04 05:43:31 crc kubenswrapper[4992]: I1004 05:43:31.319539 4992 scope.go:117] "RemoveContainer" containerID="16c8e0fd362eabdee25859fc3a115ec96ecbbc83c7ed8bdb86e06e7b4d2a45d8" Oct 04 05:43:31 crc kubenswrapper[4992]: I1004 05:43:31.785080 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" event={"ID":"d6d4bf54-7f00-4c33-830c-8368dd9678df","Type":"ContainerStarted","Data":"3d80189a2662ec16c207088ae1d6c3074b46f6bf99a98cacb5a9d6faf7f71880"} Oct 04 05:43:33 crc kubenswrapper[4992]: I1004 05:43:33.812477 4992 generic.go:334] "Generic (PLEG): container finished" podID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerID="2481c6e385e3c1770500933799ebfe1a0ae3b3478ee2dbc38ca2d149445a3bb9" exitCode=0 Oct 04 05:43:33 crc kubenswrapper[4992]: I1004 05:43:33.812572 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-df5cv/must-gather-8v9r4" event={"ID":"4cc5e9e6-5741-48ee-97aa-c694705babf5","Type":"ContainerDied","Data":"2481c6e385e3c1770500933799ebfe1a0ae3b3478ee2dbc38ca2d149445a3bb9"} Oct 04 05:43:33 crc kubenswrapper[4992]: I1004 05:43:33.813870 4992 scope.go:117] "RemoveContainer" containerID="2481c6e385e3c1770500933799ebfe1a0ae3b3478ee2dbc38ca2d149445a3bb9" Oct 04 05:43:34 crc kubenswrapper[4992]: I1004 05:43:34.082740 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-df5cv_must-gather-8v9r4_4cc5e9e6-5741-48ee-97aa-c694705babf5/gather/0.log" Oct 04 05:43:41 crc kubenswrapper[4992]: I1004 05:43:41.465834 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-df5cv/must-gather-8v9r4"] Oct 04 05:43:41 crc kubenswrapper[4992]: I1004 05:43:41.466641 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-df5cv/must-gather-8v9r4" podUID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerName="copy" containerID="cri-o://6a3292ee2b54647420b2fc86a26f0c289fe04cd9de8302eb2a2a5740015baa67" gracePeriod=2 Oct 04 05:43:41 crc kubenswrapper[4992]: I1004 05:43:41.479014 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-df5cv/must-gather-8v9r4"] Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:41.913971 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-df5cv_must-gather-8v9r4_4cc5e9e6-5741-48ee-97aa-c694705babf5/copy/0.log" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:41.914756 4992 generic.go:334] "Generic (PLEG): container finished" podID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerID="6a3292ee2b54647420b2fc86a26f0c289fe04cd9de8302eb2a2a5740015baa67" exitCode=143 Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:41.914801 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d12a84c0fda2ab1b67a06b2ab556583a61a95590d23840d93a073c5996c5a38" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:41.928512 4992 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-df5cv_must-gather-8v9r4_4cc5e9e6-5741-48ee-97aa-c694705babf5/copy/0.log" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:41.930242 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:41.969647 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lth5\" (UniqueName: \"kubernetes.io/projected/4cc5e9e6-5741-48ee-97aa-c694705babf5-kube-api-access-9lth5\") pod \"4cc5e9e6-5741-48ee-97aa-c694705babf5\" (UID: \"4cc5e9e6-5741-48ee-97aa-c694705babf5\") " Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:41.969756 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4cc5e9e6-5741-48ee-97aa-c694705babf5-must-gather-output\") pod \"4cc5e9e6-5741-48ee-97aa-c694705babf5\" (UID: \"4cc5e9e6-5741-48ee-97aa-c694705babf5\") " Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:41.977772 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc5e9e6-5741-48ee-97aa-c694705babf5-kube-api-access-9lth5" (OuterVolumeSpecName: "kube-api-access-9lth5") pod "4cc5e9e6-5741-48ee-97aa-c694705babf5" (UID: "4cc5e9e6-5741-48ee-97aa-c694705babf5"). InnerVolumeSpecName "kube-api-access-9lth5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:42.074129 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lth5\" (UniqueName: \"kubernetes.io/projected/4cc5e9e6-5741-48ee-97aa-c694705babf5-kube-api-access-9lth5\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:42.143870 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc5e9e6-5741-48ee-97aa-c694705babf5-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "4cc5e9e6-5741-48ee-97aa-c694705babf5" (UID: "4cc5e9e6-5741-48ee-97aa-c694705babf5"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:42.175579 4992 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4cc5e9e6-5741-48ee-97aa-c694705babf5-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:42.334806 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cc5e9e6-5741-48ee-97aa-c694705babf5" path="/var/lib/kubelet/pods/4cc5e9e6-5741-48ee-97aa-c694705babf5/volumes" Oct 04 05:43:42 crc kubenswrapper[4992]: I1004 05:43:42.921851 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-df5cv/must-gather-8v9r4" Oct 04 05:44:13 crc kubenswrapper[4992]: I1004 05:44:13.993973 4992 scope.go:117] "RemoveContainer" containerID="2481c6e385e3c1770500933799ebfe1a0ae3b3478ee2dbc38ca2d149445a3bb9" Oct 04 05:44:14 crc kubenswrapper[4992]: I1004 05:44:14.077984 4992 scope.go:117] "RemoveContainer" containerID="6a3292ee2b54647420b2fc86a26f0c289fe04cd9de8302eb2a2a5740015baa67" Oct 04 05:44:14 crc kubenswrapper[4992]: I1004 05:44:14.097879 4992 scope.go:117] "RemoveContainer" containerID="28ae4c509a0f46149029a591f04c1040344b5f1c58f1331dd8d0b02960c55680" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.168692 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz"] Oct 04 05:45:00 crc kubenswrapper[4992]: E1004 05:45:00.174846 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24db047-390c-4acb-9770-536264556a07" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.174865 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24db047-390c-4acb-9770-536264556a07" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4992]: E1004 05:45:00.174884 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24db047-390c-4acb-9770-536264556a07" containerName="extract-utilities" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.174890 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24db047-390c-4acb-9770-536264556a07" containerName="extract-utilities" Oct 04 05:45:00 crc kubenswrapper[4992]: E1004 05:45:00.174902 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24db047-390c-4acb-9770-536264556a07" containerName="extract-content" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.174908 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24db047-390c-4acb-9770-536264556a07" containerName="extract-content" Oct 04 05:45:00 crc kubenswrapper[4992]: E1004 05:45:00.174940 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerName="gather" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.174946 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerName="gather" Oct 04 05:45:00 crc kubenswrapper[4992]: E1004 05:45:00.174960 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerName="copy" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.174966 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerName="copy" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.175334 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerName="gather" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.175348 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc5e9e6-5741-48ee-97aa-c694705babf5" containerName="copy" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.175383 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24db047-390c-4acb-9770-536264556a07" containerName="registry-server" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.176091 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.178844 4992 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.179052 4992 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.184419 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz"] Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.210150 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4064c057-c003-4775-9fa8-a148cafa9567-secret-volume\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.210299 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx2q5\" (UniqueName: \"kubernetes.io/projected/4064c057-c003-4775-9fa8-a148cafa9567-kube-api-access-rx2q5\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.210444 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4064c057-c003-4775-9fa8-a148cafa9567-config-volume\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.311227 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4064c057-c003-4775-9fa8-a148cafa9567-config-volume\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.311651 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4064c057-c003-4775-9fa8-a148cafa9567-secret-volume\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.312167 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4064c057-c003-4775-9fa8-a148cafa9567-config-volume\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.312602 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx2q5\" (UniqueName: \"kubernetes.io/projected/4064c057-c003-4775-9fa8-a148cafa9567-kube-api-access-rx2q5\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.320142 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4064c057-c003-4775-9fa8-a148cafa9567-secret-volume\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.333147 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx2q5\" (UniqueName: \"kubernetes.io/projected/4064c057-c003-4775-9fa8-a148cafa9567-kube-api-access-rx2q5\") pod \"collect-profiles-29325945-5s8sz\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.504493 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:00 crc kubenswrapper[4992]: I1004 05:45:00.976512 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz"] Oct 04 05:45:01 crc kubenswrapper[4992]: I1004 05:45:01.744478 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" event={"ID":"4064c057-c003-4775-9fa8-a148cafa9567","Type":"ContainerStarted","Data":"073d3a88d92737289f6264d421ed53aadc9699f038296e9b9df145c214813b8f"} Oct 04 05:45:01 crc kubenswrapper[4992]: I1004 05:45:01.744799 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" event={"ID":"4064c057-c003-4775-9fa8-a148cafa9567","Type":"ContainerStarted","Data":"8610f69db8070b9cae2dead998d8f94af28ca10f75ef22dc233124d854ae53ab"} Oct 04 05:45:01 crc kubenswrapper[4992]: I1004 05:45:01.762529 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" podStartSLOduration=1.7625127919999999 podStartE2EDuration="1.762512792s" podCreationTimestamp="2025-10-04 05:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 05:45:01.761887785 +0000 UTC m=+4335.609563283" watchObservedRunningTime="2025-10-04 05:45:01.762512792 +0000 UTC m=+4335.610188270" Oct 04 05:45:02 crc kubenswrapper[4992]: I1004 05:45:02.757926 4992 generic.go:334] "Generic (PLEG): container finished" podID="4064c057-c003-4775-9fa8-a148cafa9567" containerID="073d3a88d92737289f6264d421ed53aadc9699f038296e9b9df145c214813b8f" exitCode=0 Oct 04 05:45:02 crc kubenswrapper[4992]: I1004 05:45:02.758107 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" event={"ID":"4064c057-c003-4775-9fa8-a148cafa9567","Type":"ContainerDied","Data":"073d3a88d92737289f6264d421ed53aadc9699f038296e9b9df145c214813b8f"} Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.172582 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.300179 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4064c057-c003-4775-9fa8-a148cafa9567-secret-volume\") pod \"4064c057-c003-4775-9fa8-a148cafa9567\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.300280 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx2q5\" (UniqueName: \"kubernetes.io/projected/4064c057-c003-4775-9fa8-a148cafa9567-kube-api-access-rx2q5\") pod \"4064c057-c003-4775-9fa8-a148cafa9567\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.300322 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4064c057-c003-4775-9fa8-a148cafa9567-config-volume\") pod \"4064c057-c003-4775-9fa8-a148cafa9567\" (UID: \"4064c057-c003-4775-9fa8-a148cafa9567\") " Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.301727 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4064c057-c003-4775-9fa8-a148cafa9567-config-volume" (OuterVolumeSpecName: "config-volume") pod "4064c057-c003-4775-9fa8-a148cafa9567" (UID: "4064c057-c003-4775-9fa8-a148cafa9567"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.306709 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4064c057-c003-4775-9fa8-a148cafa9567-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4064c057-c003-4775-9fa8-a148cafa9567" (UID: "4064c057-c003-4775-9fa8-a148cafa9567"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.307519 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4064c057-c003-4775-9fa8-a148cafa9567-kube-api-access-rx2q5" (OuterVolumeSpecName: "kube-api-access-rx2q5") pod "4064c057-c003-4775-9fa8-a148cafa9567" (UID: "4064c057-c003-4775-9fa8-a148cafa9567"). InnerVolumeSpecName "kube-api-access-rx2q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.403896 4992 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4064c057-c003-4775-9fa8-a148cafa9567-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.403953 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx2q5\" (UniqueName: \"kubernetes.io/projected/4064c057-c003-4775-9fa8-a148cafa9567-kube-api-access-rx2q5\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.403973 4992 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4064c057-c003-4775-9fa8-a148cafa9567-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.780824 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" event={"ID":"4064c057-c003-4775-9fa8-a148cafa9567","Type":"ContainerDied","Data":"8610f69db8070b9cae2dead998d8f94af28ca10f75ef22dc233124d854ae53ab"} Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.780870 4992 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8610f69db8070b9cae2dead998d8f94af28ca10f75ef22dc233124d854ae53ab" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.781226 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29325945-5s8sz" Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.860341 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7"] Oct 04 05:45:04 crc kubenswrapper[4992]: I1004 05:45:04.869028 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29325900-zf8n7"] Oct 04 05:45:06 crc kubenswrapper[4992]: I1004 05:45:06.329266 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9d992b6-7a89-4ebe-b5e4-b845019b26f4" path="/var/lib/kubelet/pods/f9d992b6-7a89-4ebe-b5e4-b845019b26f4/volumes" Oct 04 05:45:14 crc kubenswrapper[4992]: I1004 05:45:14.154678 4992 scope.go:117] "RemoveContainer" containerID="a057a37fcbe912bed8e156b72be425604726ffaa272d04c3ab309fb33f57c3c1" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.653747 4992 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vxjs7"] Oct 04 05:45:21 crc kubenswrapper[4992]: E1004 05:45:21.654927 4992 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4064c057-c003-4775-9fa8-a148cafa9567" containerName="collect-profiles" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.654946 4992 state_mem.go:107] "Deleted CPUSet assignment" podUID="4064c057-c003-4775-9fa8-a148cafa9567" containerName="collect-profiles" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.655335 4992 memory_manager.go:354] "RemoveStaleState removing state" podUID="4064c057-c003-4775-9fa8-a148cafa9567" containerName="collect-profiles" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.657558 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.673904 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vxjs7"] Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.783499 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whbnh\" (UniqueName: \"kubernetes.io/projected/2f405798-4110-464f-9a7a-bcbc8a296a62-kube-api-access-whbnh\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.783628 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-utilities\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.783722 4992 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-catalog-content\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.886007 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whbnh\" (UniqueName: \"kubernetes.io/projected/2f405798-4110-464f-9a7a-bcbc8a296a62-kube-api-access-whbnh\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.886122 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-utilities\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.886214 4992 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-catalog-content\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.886789 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-catalog-content\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.886863 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-utilities\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.905299 4992 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whbnh\" (UniqueName: \"kubernetes.io/projected/2f405798-4110-464f-9a7a-bcbc8a296a62-kube-api-access-whbnh\") pod \"redhat-marketplace-vxjs7\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:21 crc kubenswrapper[4992]: I1004 05:45:21.990315 4992 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:22 crc kubenswrapper[4992]: I1004 05:45:22.482111 4992 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vxjs7"] Oct 04 05:45:22 crc kubenswrapper[4992]: I1004 05:45:22.980173 4992 generic.go:334] "Generic (PLEG): container finished" podID="2f405798-4110-464f-9a7a-bcbc8a296a62" containerID="33940e2cc2e50a253c9cc8c7eb012825bc205df3cb9f6e0003325198ab0a713d" exitCode=0 Oct 04 05:45:22 crc kubenswrapper[4992]: I1004 05:45:22.980335 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxjs7" event={"ID":"2f405798-4110-464f-9a7a-bcbc8a296a62","Type":"ContainerDied","Data":"33940e2cc2e50a253c9cc8c7eb012825bc205df3cb9f6e0003325198ab0a713d"} Oct 04 05:45:22 crc kubenswrapper[4992]: I1004 05:45:22.980441 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxjs7" event={"ID":"2f405798-4110-464f-9a7a-bcbc8a296a62","Type":"ContainerStarted","Data":"dcb82b4670aeaa57a6a0b4b9b54aabbc37aee68e339b62a35325742b179e32d9"} Oct 04 05:45:22 crc kubenswrapper[4992]: I1004 05:45:22.982103 4992 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 05:45:26 crc kubenswrapper[4992]: I1004 05:45:26.025554 4992 generic.go:334] "Generic (PLEG): container finished" podID="2f405798-4110-464f-9a7a-bcbc8a296a62" containerID="993ae56ab8b748b63825d30b08381244975fcfdab9e5c52214983b54bac5b583" exitCode=0 Oct 04 05:45:26 crc kubenswrapper[4992]: I1004 05:45:26.025650 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxjs7" event={"ID":"2f405798-4110-464f-9a7a-bcbc8a296a62","Type":"ContainerDied","Data":"993ae56ab8b748b63825d30b08381244975fcfdab9e5c52214983b54bac5b583"} Oct 04 05:45:28 crc kubenswrapper[4992]: I1004 05:45:28.047370 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxjs7" event={"ID":"2f405798-4110-464f-9a7a-bcbc8a296a62","Type":"ContainerStarted","Data":"68266db4fd8c4af3d2421501401071d8c00ba4f050ba2c190a1de9ffca268200"} Oct 04 05:45:28 crc kubenswrapper[4992]: I1004 05:45:28.069982 4992 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vxjs7" podStartSLOduration=2.809130291 podStartE2EDuration="7.069961199s" podCreationTimestamp="2025-10-04 05:45:21 +0000 UTC" firstStartedPulling="2025-10-04 05:45:22.981844495 +0000 UTC m=+4356.829519963" lastFinishedPulling="2025-10-04 05:45:27.242675393 +0000 UTC m=+4361.090350871" observedRunningTime="2025-10-04 05:45:28.064227146 +0000 UTC m=+4361.911902624" watchObservedRunningTime="2025-10-04 05:45:28.069961199 +0000 UTC m=+4361.917636667" Oct 04 05:45:31 crc kubenswrapper[4992]: I1004 05:45:31.990760 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:31 crc kubenswrapper[4992]: I1004 05:45:31.991449 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:32 crc kubenswrapper[4992]: I1004 05:45:32.075991 4992 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:32 crc kubenswrapper[4992]: I1004 05:45:32.164607 4992 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:32 crc kubenswrapper[4992]: I1004 05:45:32.313833 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vxjs7"] Oct 04 05:45:34 crc kubenswrapper[4992]: I1004 05:45:34.118123 4992 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vxjs7" podUID="2f405798-4110-464f-9a7a-bcbc8a296a62" containerName="registry-server" containerID="cri-o://68266db4fd8c4af3d2421501401071d8c00ba4f050ba2c190a1de9ffca268200" gracePeriod=2 Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.132320 4992 generic.go:334] "Generic (PLEG): container finished" podID="2f405798-4110-464f-9a7a-bcbc8a296a62" containerID="68266db4fd8c4af3d2421501401071d8c00ba4f050ba2c190a1de9ffca268200" exitCode=0 Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.132443 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxjs7" event={"ID":"2f405798-4110-464f-9a7a-bcbc8a296a62","Type":"ContainerDied","Data":"68266db4fd8c4af3d2421501401071d8c00ba4f050ba2c190a1de9ffca268200"} Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.335049 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.467016 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-catalog-content\") pod \"2f405798-4110-464f-9a7a-bcbc8a296a62\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.467249 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whbnh\" (UniqueName: \"kubernetes.io/projected/2f405798-4110-464f-9a7a-bcbc8a296a62-kube-api-access-whbnh\") pod \"2f405798-4110-464f-9a7a-bcbc8a296a62\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.467396 4992 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-utilities\") pod \"2f405798-4110-464f-9a7a-bcbc8a296a62\" (UID: \"2f405798-4110-464f-9a7a-bcbc8a296a62\") " Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.469158 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-utilities" (OuterVolumeSpecName: "utilities") pod "2f405798-4110-464f-9a7a-bcbc8a296a62" (UID: "2f405798-4110-464f-9a7a-bcbc8a296a62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.476281 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f405798-4110-464f-9a7a-bcbc8a296a62-kube-api-access-whbnh" (OuterVolumeSpecName: "kube-api-access-whbnh") pod "2f405798-4110-464f-9a7a-bcbc8a296a62" (UID: "2f405798-4110-464f-9a7a-bcbc8a296a62"). InnerVolumeSpecName "kube-api-access-whbnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.487327 4992 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f405798-4110-464f-9a7a-bcbc8a296a62" (UID: "2f405798-4110-464f-9a7a-bcbc8a296a62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.569665 4992 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whbnh\" (UniqueName: \"kubernetes.io/projected/2f405798-4110-464f-9a7a-bcbc8a296a62-kube-api-access-whbnh\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.569701 4992 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:35 crc kubenswrapper[4992]: I1004 05:45:35.569710 4992 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f405798-4110-464f-9a7a-bcbc8a296a62-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 05:45:36 crc kubenswrapper[4992]: I1004 05:45:36.146096 4992 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vxjs7" event={"ID":"2f405798-4110-464f-9a7a-bcbc8a296a62","Type":"ContainerDied","Data":"dcb82b4670aeaa57a6a0b4b9b54aabbc37aee68e339b62a35325742b179e32d9"} Oct 04 05:45:36 crc kubenswrapper[4992]: I1004 05:45:36.146574 4992 scope.go:117] "RemoveContainer" containerID="68266db4fd8c4af3d2421501401071d8c00ba4f050ba2c190a1de9ffca268200" Oct 04 05:45:36 crc kubenswrapper[4992]: I1004 05:45:36.146197 4992 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vxjs7" Oct 04 05:45:36 crc kubenswrapper[4992]: I1004 05:45:36.191128 4992 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vxjs7"] Oct 04 05:45:36 crc kubenswrapper[4992]: I1004 05:45:36.194514 4992 scope.go:117] "RemoveContainer" containerID="993ae56ab8b748b63825d30b08381244975fcfdab9e5c52214983b54bac5b583" Oct 04 05:45:36 crc kubenswrapper[4992]: I1004 05:45:36.200487 4992 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vxjs7"] Oct 04 05:45:36 crc kubenswrapper[4992]: I1004 05:45:36.223463 4992 scope.go:117] "RemoveContainer" containerID="33940e2cc2e50a253c9cc8c7eb012825bc205df3cb9f6e0003325198ab0a713d" Oct 04 05:45:36 crc kubenswrapper[4992]: I1004 05:45:36.339124 4992 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f405798-4110-464f-9a7a-bcbc8a296a62" path="/var/lib/kubelet/pods/2f405798-4110-464f-9a7a-bcbc8a296a62/volumes" Oct 04 05:45:50 crc kubenswrapper[4992]: I1004 05:45:50.894201 4992 patch_prober.go:28] interesting pod/machine-config-daemon-p5l9q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 05:45:50 crc kubenswrapper[4992]: I1004 05:45:50.894700 4992 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p5l9q" podUID="d6d4bf54-7f00-4c33-830c-8368dd9678df" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070132246024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070132247017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070121225016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070121225015450 5ustar corecore